00:00:00.001 Started by upstream project "autotest-per-patch" build number 132534 00:00:00.001 originally caused by: 00:00:00.001 Started by user sys_sgci 00:00:00.049 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-phy.groovy 00:00:00.050 The recommended git tool is: git 00:00:00.050 using credential 00000000-0000-0000-0000-000000000002 00:00:00.052 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.080 Fetching changes from the remote Git repository 00:00:00.083 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.138 Using shallow fetch with depth 1 00:00:00.138 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.138 > git --version # timeout=10 00:00:00.203 > git --version # 'git version 2.39.2' 00:00:00.203 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.249 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.249 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:04.820 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:04.830 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:04.843 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:04.843 > git config core.sparsecheckout # timeout=10 00:00:04.854 > git read-tree -mu HEAD # timeout=10 00:00:04.871 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:04.893 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:04.893 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:04.970 [Pipeline] Start of Pipeline 00:00:04.979 [Pipeline] library 00:00:04.981 Loading library shm_lib@master 00:00:04.981 Library shm_lib@master is cached. Copying from home. 00:00:04.993 [Pipeline] node 00:00:05.000 Running on CYP12 in /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:05.002 [Pipeline] { 00:00:05.009 [Pipeline] catchError 00:00:05.010 [Pipeline] { 00:00:05.019 [Pipeline] wrap 00:00:05.026 [Pipeline] { 00:00:05.031 [Pipeline] stage 00:00:05.032 [Pipeline] { (Prologue) 00:00:05.220 [Pipeline] sh 00:00:05.504 + logger -p user.info -t JENKINS-CI 00:00:05.517 [Pipeline] echo 00:00:05.518 Node: CYP12 00:00:05.523 [Pipeline] sh 00:00:05.820 [Pipeline] setCustomBuildProperty 00:00:05.830 [Pipeline] echo 00:00:05.831 Cleanup processes 00:00:05.835 [Pipeline] sh 00:00:06.116 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.116 3510754 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.128 [Pipeline] sh 00:00:06.409 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.409 ++ grep -v 'sudo pgrep' 00:00:06.409 ++ awk '{print $1}' 00:00:06.409 + sudo kill -9 00:00:06.409 + true 00:00:06.423 [Pipeline] cleanWs 00:00:06.433 [WS-CLEANUP] Deleting project workspace... 00:00:06.433 [WS-CLEANUP] Deferred wipeout is used... 00:00:06.441 [WS-CLEANUP] done 00:00:06.444 [Pipeline] setCustomBuildProperty 00:00:06.486 [Pipeline] sh 00:00:06.769 + sudo git config --global --replace-all safe.directory '*' 00:00:06.849 [Pipeline] httpRequest 00:00:07.440 [Pipeline] echo 00:00:07.442 Sorcerer 10.211.164.101 is alive 00:00:07.450 [Pipeline] retry 00:00:07.451 [Pipeline] { 00:00:07.462 [Pipeline] httpRequest 00:00:07.469 HttpMethod: GET 00:00:07.470 URL: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:07.470 Sending request to url: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:07.473 Response Code: HTTP/1.1 200 OK 00:00:07.473 Success: Status code 200 is in the accepted range: 200,404 00:00:07.473 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:09.018 [Pipeline] } 00:00:09.036 [Pipeline] // retry 00:00:09.044 [Pipeline] sh 00:00:09.330 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:09.353 [Pipeline] httpRequest 00:00:09.758 [Pipeline] echo 00:00:09.760 Sorcerer 10.211.164.101 is alive 00:00:09.770 [Pipeline] retry 00:00:09.772 [Pipeline] { 00:00:09.788 [Pipeline] httpRequest 00:00:09.793 HttpMethod: GET 00:00:09.794 URL: http://10.211.164.101/packages/spdk_bb877d8c1896828625e90fadfd5df38b20e54397.tar.gz 00:00:09.794 Sending request to url: http://10.211.164.101/packages/spdk_bb877d8c1896828625e90fadfd5df38b20e54397.tar.gz 00:00:09.809 Response Code: HTTP/1.1 200 OK 00:00:09.809 Success: Status code 200 is in the accepted range: 200,404 00:00:09.810 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk_bb877d8c1896828625e90fadfd5df38b20e54397.tar.gz 00:01:13.920 [Pipeline] } 00:01:13.941 [Pipeline] // retry 00:01:13.950 [Pipeline] sh 00:01:14.238 + tar --no-same-owner -xf spdk_bb877d8c1896828625e90fadfd5df38b20e54397.tar.gz 00:01:17.549 [Pipeline] sh 00:01:17.835 + git -C spdk log --oneline -n5 00:01:17.835 bb877d8c1 nvmf: Expose DIF type of namespace to host again 00:01:17.835 9f3071c5f nvmf: Set bdev_ext_io_opts::dif_check_flags_exclude_mask for read/write 00:01:17.835 5ca6db5da nvme_spec: Add SPDK_NVME_IO_FLAGS_PRCHK_MASK 00:01:17.835 f7ce15267 bdev: Insert or overwrite metadata using bounce/accel buffer if NVMe PRACT is set 00:01:17.835 aa58c9e0b dif: Add spdk_dif_pi_format_get_size() to use for NVMe PRACT 00:01:17.848 [Pipeline] } 00:01:17.862 [Pipeline] // stage 00:01:17.872 [Pipeline] stage 00:01:17.874 [Pipeline] { (Prepare) 00:01:17.892 [Pipeline] writeFile 00:01:17.908 [Pipeline] sh 00:01:18.194 + logger -p user.info -t JENKINS-CI 00:01:18.207 [Pipeline] sh 00:01:18.496 + logger -p user.info -t JENKINS-CI 00:01:18.511 [Pipeline] sh 00:01:18.796 + cat autorun-spdk.conf 00:01:18.796 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:18.796 SPDK_TEST_NVMF=1 00:01:18.796 SPDK_TEST_NVME_CLI=1 00:01:18.796 SPDK_TEST_NVMF_TRANSPORT=tcp 00:01:18.796 SPDK_TEST_NVMF_NICS=e810 00:01:18.796 SPDK_TEST_VFIOUSER=1 00:01:18.796 SPDK_RUN_UBSAN=1 00:01:18.796 NET_TYPE=phy 00:01:18.804 RUN_NIGHTLY=0 00:01:18.809 [Pipeline] readFile 00:01:18.832 [Pipeline] withEnv 00:01:18.834 [Pipeline] { 00:01:18.845 [Pipeline] sh 00:01:19.132 + set -ex 00:01:19.132 + [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf ]] 00:01:19.132 + source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:01:19.132 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:19.132 ++ SPDK_TEST_NVMF=1 00:01:19.132 ++ SPDK_TEST_NVME_CLI=1 00:01:19.132 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:01:19.132 ++ SPDK_TEST_NVMF_NICS=e810 00:01:19.132 ++ SPDK_TEST_VFIOUSER=1 00:01:19.132 ++ SPDK_RUN_UBSAN=1 00:01:19.132 ++ NET_TYPE=phy 00:01:19.132 ++ RUN_NIGHTLY=0 00:01:19.132 + case $SPDK_TEST_NVMF_NICS in 00:01:19.132 + DRIVERS=ice 00:01:19.132 + [[ tcp == \r\d\m\a ]] 00:01:19.132 + [[ -n ice ]] 00:01:19.132 + sudo rmmod mlx4_ib mlx5_ib irdma i40iw iw_cxgb4 00:01:19.132 rmmod: ERROR: Module mlx4_ib is not currently loaded 00:01:19.132 rmmod: ERROR: Module mlx5_ib is not currently loaded 00:01:19.132 rmmod: ERROR: Module irdma is not currently loaded 00:01:19.132 rmmod: ERROR: Module i40iw is not currently loaded 00:01:19.132 rmmod: ERROR: Module iw_cxgb4 is not currently loaded 00:01:19.132 + true 00:01:19.132 + for D in $DRIVERS 00:01:19.132 + sudo modprobe ice 00:01:19.132 + exit 0 00:01:19.142 [Pipeline] } 00:01:19.157 [Pipeline] // withEnv 00:01:19.164 [Pipeline] } 00:01:19.179 [Pipeline] // stage 00:01:19.189 [Pipeline] catchError 00:01:19.191 [Pipeline] { 00:01:19.206 [Pipeline] timeout 00:01:19.206 Timeout set to expire in 1 hr 0 min 00:01:19.208 [Pipeline] { 00:01:19.222 [Pipeline] stage 00:01:19.224 [Pipeline] { (Tests) 00:01:19.240 [Pipeline] sh 00:01:19.530 + jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:01:19.530 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:01:19.530 + DIR_ROOT=/var/jenkins/workspace/nvmf-tcp-phy-autotest 00:01:19.530 + [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest ]] 00:01:19.530 + DIR_SPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:01:19.530 + DIR_OUTPUT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:01:19.530 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk ]] 00:01:19.530 + [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:01:19.530 + mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:01:19.530 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:01:19.530 + [[ nvmf-tcp-phy-autotest == pkgdep-* ]] 00:01:19.530 + cd /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:01:19.530 + source /etc/os-release 00:01:19.530 ++ NAME='Fedora Linux' 00:01:19.530 ++ VERSION='39 (Cloud Edition)' 00:01:19.530 ++ ID=fedora 00:01:19.530 ++ VERSION_ID=39 00:01:19.530 ++ VERSION_CODENAME= 00:01:19.530 ++ PLATFORM_ID=platform:f39 00:01:19.530 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:01:19.530 ++ ANSI_COLOR='0;38;2;60;110;180' 00:01:19.530 ++ LOGO=fedora-logo-icon 00:01:19.530 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:01:19.530 ++ HOME_URL=https://fedoraproject.org/ 00:01:19.530 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:01:19.530 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:01:19.530 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:01:19.530 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:01:19.530 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:01:19.530 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:01:19.530 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:01:19.530 ++ SUPPORT_END=2024-11-12 00:01:19.530 ++ VARIANT='Cloud Edition' 00:01:19.530 ++ VARIANT_ID=cloud 00:01:19.530 + uname -a 00:01:19.530 Linux spdk-cyp-12 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:01:19.530 + sudo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:01:22.831 Hugepages 00:01:22.831 node hugesize free / total 00:01:22.831 node0 1048576kB 0 / 0 00:01:22.831 node0 2048kB 0 / 0 00:01:22.831 node1 1048576kB 0 / 0 00:01:22.831 node1 2048kB 0 / 0 00:01:22.831 00:01:22.831 Type BDF Vendor Device NUMA Driver Device Block devices 00:01:22.831 I/OAT 0000:00:01.0 8086 0b00 0 ioatdma - - 00:01:22.831 I/OAT 0000:00:01.1 8086 0b00 0 ioatdma - - 00:01:22.831 I/OAT 0000:00:01.2 8086 0b00 0 ioatdma - - 00:01:22.831 I/OAT 0000:00:01.3 8086 0b00 0 ioatdma - - 00:01:22.831 I/OAT 0000:00:01.4 8086 0b00 0 ioatdma - - 00:01:22.831 I/OAT 0000:00:01.5 8086 0b00 0 ioatdma - - 00:01:22.831 I/OAT 0000:00:01.6 8086 0b00 0 ioatdma - - 00:01:22.831 I/OAT 0000:00:01.7 8086 0b00 0 ioatdma - - 00:01:23.093 NVMe 0000:65:00.0 144d a80a 0 nvme nvme0 nvme0n1 00:01:23.093 I/OAT 0000:80:01.0 8086 0b00 1 ioatdma - - 00:01:23.093 I/OAT 0000:80:01.1 8086 0b00 1 ioatdma - - 00:01:23.093 I/OAT 0000:80:01.2 8086 0b00 1 ioatdma - - 00:01:23.093 I/OAT 0000:80:01.3 8086 0b00 1 ioatdma - - 00:01:23.093 I/OAT 0000:80:01.4 8086 0b00 1 ioatdma - - 00:01:23.093 I/OAT 0000:80:01.5 8086 0b00 1 ioatdma - - 00:01:23.093 I/OAT 0000:80:01.6 8086 0b00 1 ioatdma - - 00:01:23.093 I/OAT 0000:80:01.7 8086 0b00 1 ioatdma - - 00:01:23.093 + rm -f /tmp/spdk-ld-path 00:01:23.093 + source autorun-spdk.conf 00:01:23.093 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:23.093 ++ SPDK_TEST_NVMF=1 00:01:23.093 ++ SPDK_TEST_NVME_CLI=1 00:01:23.093 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:01:23.093 ++ SPDK_TEST_NVMF_NICS=e810 00:01:23.093 ++ SPDK_TEST_VFIOUSER=1 00:01:23.093 ++ SPDK_RUN_UBSAN=1 00:01:23.093 ++ NET_TYPE=phy 00:01:23.093 ++ RUN_NIGHTLY=0 00:01:23.093 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:01:23.093 + [[ -n '' ]] 00:01:23.093 + sudo git config --global --add safe.directory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:01:23.093 + for M in /var/spdk/build-*-manifest.txt 00:01:23.093 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:01:23.093 + cp /var/spdk/build-kernel-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:01:23.093 + for M in /var/spdk/build-*-manifest.txt 00:01:23.093 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:01:23.093 + cp /var/spdk/build-pkg-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:01:23.093 + for M in /var/spdk/build-*-manifest.txt 00:01:23.093 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:01:23.093 + cp /var/spdk/build-repo-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:01:23.093 ++ uname 00:01:23.093 + [[ Linux == \L\i\n\u\x ]] 00:01:23.093 + sudo dmesg -T 00:01:23.093 + sudo dmesg --clear 00:01:23.355 + dmesg_pid=3512425 00:01:23.355 + [[ Fedora Linux == FreeBSD ]] 00:01:23.355 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:01:23.355 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:01:23.355 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:01:23.355 + [[ -x /usr/src/fio-static/fio ]] 00:01:23.355 + export FIO_BIN=/usr/src/fio-static/fio 00:01:23.355 + FIO_BIN=/usr/src/fio-static/fio 00:01:23.355 + sudo dmesg -Tw 00:01:23.355 + [[ '' == \/\v\a\r\/\j\e\n\k\i\n\s\/\w\o\r\k\s\p\a\c\e\/\n\v\m\f\-\t\c\p\-\p\h\y\-\a\u\t\o\t\e\s\t\/\q\e\m\u\_\v\f\i\o\/* ]] 00:01:23.355 + [[ ! -v VFIO_QEMU_BIN ]] 00:01:23.355 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:01:23.355 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:01:23.355 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:01:23.355 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:01:23.355 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:01:23.355 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:01:23.355 + spdk/autorun.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:01:23.355 18:57:35 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:01:23.355 18:57:35 -- spdk/autorun.sh@20 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@2 -- $ SPDK_TEST_NVMF=1 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@3 -- $ SPDK_TEST_NVME_CLI=1 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@4 -- $ SPDK_TEST_NVMF_TRANSPORT=tcp 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@5 -- $ SPDK_TEST_NVMF_NICS=e810 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@6 -- $ SPDK_TEST_VFIOUSER=1 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@7 -- $ SPDK_RUN_UBSAN=1 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@8 -- $ NET_TYPE=phy 00:01:23.355 18:57:35 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@9 -- $ RUN_NIGHTLY=0 00:01:23.355 18:57:35 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:01:23.355 18:57:35 -- spdk/autorun.sh@25 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autobuild.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:01:23.355 18:57:35 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:01:23.355 18:57:35 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:01:23.355 18:57:35 -- scripts/common.sh@15 -- $ shopt -s extglob 00:01:23.355 18:57:35 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:01:23.355 18:57:35 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:01:23.355 18:57:35 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:01:23.355 18:57:35 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:23.355 18:57:35 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:23.355 18:57:35 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:23.355 18:57:35 -- paths/export.sh@5 -- $ export PATH 00:01:23.355 18:57:35 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:23.355 18:57:35 -- common/autobuild_common.sh@492 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:01:23.355 18:57:35 -- common/autobuild_common.sh@493 -- $ date +%s 00:01:23.355 18:57:35 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1732643855.XXXXXX 00:01:23.355 18:57:35 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1732643855.ajWG5M 00:01:23.355 18:57:35 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:01:23.355 18:57:35 -- common/autobuild_common.sh@499 -- $ '[' -n '' ']' 00:01:23.355 18:57:35 -- common/autobuild_common.sh@502 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:01:23.355 18:57:35 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:01:23.355 18:57:35 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:01:23.355 18:57:35 -- common/autobuild_common.sh@509 -- $ get_config_params 00:01:23.355 18:57:35 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:01:23.355 18:57:35 -- common/autotest_common.sh@10 -- $ set +x 00:01:23.355 18:57:35 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:01:23.355 18:57:35 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:01:23.355 18:57:35 -- pm/common@17 -- $ local monitor 00:01:23.355 18:57:35 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:23.355 18:57:35 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:23.355 18:57:35 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:23.355 18:57:35 -- pm/common@21 -- $ date +%s 00:01:23.355 18:57:35 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:23.355 18:57:35 -- pm/common@25 -- $ sleep 1 00:01:23.355 18:57:35 -- pm/common@21 -- $ date +%s 00:01:23.355 18:57:35 -- pm/common@21 -- $ date +%s 00:01:23.355 18:57:35 -- pm/common@21 -- $ date +%s 00:01:23.355 18:57:35 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1732643855 00:01:23.355 18:57:35 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1732643855 00:01:23.355 18:57:35 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1732643855 00:01:23.355 18:57:35 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1732643855 00:01:23.355 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1732643855_collect-cpu-load.pm.log 00:01:23.355 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1732643855_collect-vmstat.pm.log 00:01:23.355 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1732643855_collect-cpu-temp.pm.log 00:01:23.617 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1732643855_collect-bmc-pm.bmc.pm.log 00:01:24.557 18:57:36 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:01:24.557 18:57:36 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:01:24.557 18:57:36 -- spdk/autobuild.sh@12 -- $ umask 022 00:01:24.557 18:57:36 -- spdk/autobuild.sh@13 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:01:24.557 18:57:36 -- spdk/autobuild.sh@16 -- $ date -u 00:01:24.557 Tue Nov 26 05:57:36 PM UTC 2024 00:01:24.557 18:57:36 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:01:24.557 v25.01-pre-271-gbb877d8c1 00:01:24.557 18:57:36 -- spdk/autobuild.sh@19 -- $ '[' 0 -eq 1 ']' 00:01:24.557 18:57:36 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:01:24.558 18:57:36 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:01:24.558 18:57:36 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:01:24.558 18:57:36 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:01:24.558 18:57:36 -- common/autotest_common.sh@10 -- $ set +x 00:01:24.558 ************************************ 00:01:24.558 START TEST ubsan 00:01:24.558 ************************************ 00:01:24.558 18:57:36 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:01:24.558 using ubsan 00:01:24.558 00:01:24.558 real 0m0.001s 00:01:24.558 user 0m0.000s 00:01:24.558 sys 0m0.000s 00:01:24.558 18:57:36 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:01:24.558 18:57:36 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:01:24.558 ************************************ 00:01:24.558 END TEST ubsan 00:01:24.558 ************************************ 00:01:24.558 18:57:37 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:01:24.558 18:57:37 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:01:24.558 18:57:37 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:01:24.558 18:57:37 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:01:24.558 18:57:37 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:01:24.558 18:57:37 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:01:24.558 18:57:37 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:01:24.558 18:57:37 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:01:24.558 18:57:37 -- spdk/autobuild.sh@67 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user --with-shared 00:01:24.819 Using default SPDK env in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:01:24.819 Using default DPDK in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:01:25.080 Using 'verbs' RDMA provider 00:01:40.939 Configuring ISA-L (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal.log)...done. 00:01:53.171 Configuring ISA-L-crypto (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal-crypto.log)...done. 00:01:53.171 Creating mk/config.mk...done. 00:01:53.171 Creating mk/cc.flags.mk...done. 00:01:53.172 Type 'make' to build. 00:01:53.172 18:58:05 -- spdk/autobuild.sh@70 -- $ run_test make make -j144 00:01:53.172 18:58:05 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:01:53.172 18:58:05 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:01:53.172 18:58:05 -- common/autotest_common.sh@10 -- $ set +x 00:01:53.172 ************************************ 00:01:53.172 START TEST make 00:01:53.172 ************************************ 00:01:53.172 18:58:05 make -- common/autotest_common.sh@1129 -- $ make -j144 00:01:53.433 make[1]: Nothing to be done for 'all'. 00:01:54.814 The Meson build system 00:01:54.814 Version: 1.5.0 00:01:54.814 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user 00:01:54.814 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:54.814 Build type: native build 00:01:54.814 Project name: libvfio-user 00:01:54.814 Project version: 0.0.1 00:01:54.814 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:01:54.814 C linker for the host machine: cc ld.bfd 2.40-14 00:01:54.814 Host machine cpu family: x86_64 00:01:54.814 Host machine cpu: x86_64 00:01:54.814 Run-time dependency threads found: YES 00:01:54.814 Library dl found: YES 00:01:54.814 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:01:54.814 Run-time dependency json-c found: YES 0.17 00:01:54.814 Run-time dependency cmocka found: YES 1.1.7 00:01:54.814 Program pytest-3 found: NO 00:01:54.814 Program flake8 found: NO 00:01:54.814 Program misspell-fixer found: NO 00:01:54.814 Program restructuredtext-lint found: NO 00:01:54.814 Program valgrind found: YES (/usr/bin/valgrind) 00:01:54.814 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:01:54.814 Compiler for C supports arguments -Wmissing-declarations: YES 00:01:54.814 Compiler for C supports arguments -Wwrite-strings: YES 00:01:54.814 ../libvfio-user/test/meson.build:20: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:01:54.814 Program test-lspci.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-lspci.sh) 00:01:54.814 Program test-linkage.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-linkage.sh) 00:01:54.814 ../libvfio-user/test/py/meson.build:16: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:01:54.814 Build targets in project: 8 00:01:54.814 WARNING: Project specifies a minimum meson_version '>= 0.53.0' but uses features which were added in newer versions: 00:01:54.814 * 0.57.0: {'exclude_suites arg in add_test_setup'} 00:01:54.814 00:01:54.814 libvfio-user 0.0.1 00:01:54.814 00:01:54.814 User defined options 00:01:54.814 buildtype : debug 00:01:54.814 default_library: shared 00:01:54.814 libdir : /usr/local/lib 00:01:54.814 00:01:54.814 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:01:55.073 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:01:55.073 [1/37] Compiling C object samples/lspci.p/lspci.c.o 00:01:55.073 [2/37] Compiling C object samples/shadow_ioeventfd_server.p/shadow_ioeventfd_server.c.o 00:01:55.073 [3/37] Compiling C object samples/null.p/null.c.o 00:01:55.073 [4/37] Compiling C object lib/libvfio-user.so.0.0.1.p/migration.c.o 00:01:55.073 [5/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran.c.o 00:01:55.073 [6/37] Compiling C object samples/gpio-pci-idio-16.p/gpio-pci-idio-16.c.o 00:01:55.073 [7/37] Compiling C object samples/client.p/.._lib_migration.c.o 00:01:55.073 [8/37] Compiling C object test/unit_tests.p/.._lib_tran.c.o 00:01:55.073 [9/37] Compiling C object samples/client.p/.._lib_tran.c.o 00:01:55.073 [10/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci.c.o 00:01:55.073 [11/37] Compiling C object test/unit_tests.p/.._lib_irq.c.o 00:01:55.073 [12/37] Compiling C object test/unit_tests.p/mocks.c.o 00:01:55.073 [13/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci_caps.c.o 00:01:55.073 [14/37] Compiling C object lib/libvfio-user.so.0.0.1.p/irq.c.o 00:01:55.073 [15/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran_sock.c.o 00:01:55.073 [16/37] Compiling C object test/unit_tests.p/unit-tests.c.o 00:01:55.074 [17/37] Compiling C object samples/server.p/server.c.o 00:01:55.074 [18/37] Compiling C object samples/client.p/.._lib_tran_sock.c.o 00:01:55.074 [19/37] Compiling C object test/unit_tests.p/.._lib_migration.c.o 00:01:55.074 [20/37] Compiling C object test/unit_tests.p/.._lib_tran_pipe.c.o 00:01:55.074 [21/37] Compiling C object test/unit_tests.p/.._lib_dma.c.o 00:01:55.074 [22/37] Compiling C object test/unit_tests.p/.._lib_pci.c.o 00:01:55.074 [23/37] Compiling C object lib/libvfio-user.so.0.0.1.p/dma.c.o 00:01:55.074 [24/37] Compiling C object test/unit_tests.p/.._lib_tran_sock.c.o 00:01:55.074 [25/37] Compiling C object samples/client.p/client.c.o 00:01:55.074 [26/37] Compiling C object test/unit_tests.p/.._lib_pci_caps.c.o 00:01:55.334 [27/37] Linking target samples/client 00:01:55.334 [28/37] Compiling C object test/unit_tests.p/.._lib_libvfio-user.c.o 00:01:55.334 [29/37] Compiling C object lib/libvfio-user.so.0.0.1.p/libvfio-user.c.o 00:01:55.334 [30/37] Linking target lib/libvfio-user.so.0.0.1 00:01:55.334 [31/37] Linking target test/unit_tests 00:01:55.334 [32/37] Generating symbol file lib/libvfio-user.so.0.0.1.p/libvfio-user.so.0.0.1.symbols 00:01:55.596 [33/37] Linking target samples/shadow_ioeventfd_server 00:01:55.596 [34/37] Linking target samples/server 00:01:55.596 [35/37] Linking target samples/null 00:01:55.596 [36/37] Linking target samples/gpio-pci-idio-16 00:01:55.596 [37/37] Linking target samples/lspci 00:01:55.596 INFO: autodetecting backend as ninja 00:01:55.596 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:55.596 DESTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user meson install --quiet -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:55.856 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:01:55.856 ninja: no work to do. 00:02:02.444 The Meson build system 00:02:02.444 Version: 1.5.0 00:02:02.444 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk 00:02:02.444 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp 00:02:02.444 Build type: native build 00:02:02.444 Program cat found: YES (/usr/bin/cat) 00:02:02.444 Project name: DPDK 00:02:02.444 Project version: 24.03.0 00:02:02.444 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:02.444 C linker for the host machine: cc ld.bfd 2.40-14 00:02:02.444 Host machine cpu family: x86_64 00:02:02.444 Host machine cpu: x86_64 00:02:02.444 Message: ## Building in Developer Mode ## 00:02:02.444 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:02.444 Program check-symbols.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/check-symbols.sh) 00:02:02.444 Program options-ibverbs-static.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:02:02.444 Program python3 found: YES (/usr/bin/python3) 00:02:02.444 Program cat found: YES (/usr/bin/cat) 00:02:02.444 Compiler for C supports arguments -march=native: YES 00:02:02.444 Checking for size of "void *" : 8 00:02:02.444 Checking for size of "void *" : 8 (cached) 00:02:02.444 Compiler for C supports link arguments -Wl,--undefined-version: YES 00:02:02.444 Library m found: YES 00:02:02.444 Library numa found: YES 00:02:02.444 Has header "numaif.h" : YES 00:02:02.444 Library fdt found: NO 00:02:02.444 Library execinfo found: NO 00:02:02.444 Has header "execinfo.h" : YES 00:02:02.444 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:02.444 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:02.444 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:02.444 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:02.444 Run-time dependency openssl found: YES 3.1.1 00:02:02.444 Run-time dependency libpcap found: YES 1.10.4 00:02:02.444 Has header "pcap.h" with dependency libpcap: YES 00:02:02.444 Compiler for C supports arguments -Wcast-qual: YES 00:02:02.444 Compiler for C supports arguments -Wdeprecated: YES 00:02:02.444 Compiler for C supports arguments -Wformat: YES 00:02:02.444 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:02.444 Compiler for C supports arguments -Wformat-security: NO 00:02:02.444 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:02.444 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:02.444 Compiler for C supports arguments -Wnested-externs: YES 00:02:02.444 Compiler for C supports arguments -Wold-style-definition: YES 00:02:02.444 Compiler for C supports arguments -Wpointer-arith: YES 00:02:02.444 Compiler for C supports arguments -Wsign-compare: YES 00:02:02.444 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:02.444 Compiler for C supports arguments -Wundef: YES 00:02:02.444 Compiler for C supports arguments -Wwrite-strings: YES 00:02:02.444 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:02.444 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:02.444 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:02.444 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:02.444 Program objdump found: YES (/usr/bin/objdump) 00:02:02.444 Compiler for C supports arguments -mavx512f: YES 00:02:02.444 Checking if "AVX512 checking" compiles: YES 00:02:02.444 Fetching value of define "__SSE4_2__" : 1 00:02:02.444 Fetching value of define "__AES__" : 1 00:02:02.444 Fetching value of define "__AVX__" : 1 00:02:02.444 Fetching value of define "__AVX2__" : 1 00:02:02.444 Fetching value of define "__AVX512BW__" : 1 00:02:02.444 Fetching value of define "__AVX512CD__" : 1 00:02:02.444 Fetching value of define "__AVX512DQ__" : 1 00:02:02.444 Fetching value of define "__AVX512F__" : 1 00:02:02.444 Fetching value of define "__AVX512VL__" : 1 00:02:02.444 Fetching value of define "__PCLMUL__" : 1 00:02:02.444 Fetching value of define "__RDRND__" : 1 00:02:02.444 Fetching value of define "__RDSEED__" : 1 00:02:02.444 Fetching value of define "__VPCLMULQDQ__" : 1 00:02:02.444 Fetching value of define "__znver1__" : (undefined) 00:02:02.444 Fetching value of define "__znver2__" : (undefined) 00:02:02.444 Fetching value of define "__znver3__" : (undefined) 00:02:02.444 Fetching value of define "__znver4__" : (undefined) 00:02:02.444 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:02.444 Message: lib/log: Defining dependency "log" 00:02:02.444 Message: lib/kvargs: Defining dependency "kvargs" 00:02:02.444 Message: lib/telemetry: Defining dependency "telemetry" 00:02:02.444 Checking for function "getentropy" : NO 00:02:02.444 Message: lib/eal: Defining dependency "eal" 00:02:02.444 Message: lib/ring: Defining dependency "ring" 00:02:02.444 Message: lib/rcu: Defining dependency "rcu" 00:02:02.444 Message: lib/mempool: Defining dependency "mempool" 00:02:02.444 Message: lib/mbuf: Defining dependency "mbuf" 00:02:02.444 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:02.444 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:02.444 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:02.444 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:02.444 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:02.444 Fetching value of define "__VPCLMULQDQ__" : 1 (cached) 00:02:02.444 Compiler for C supports arguments -mpclmul: YES 00:02:02.444 Compiler for C supports arguments -maes: YES 00:02:02.444 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:02.444 Compiler for C supports arguments -mavx512bw: YES 00:02:02.444 Compiler for C supports arguments -mavx512dq: YES 00:02:02.444 Compiler for C supports arguments -mavx512vl: YES 00:02:02.444 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:02.444 Compiler for C supports arguments -mavx2: YES 00:02:02.444 Compiler for C supports arguments -mavx: YES 00:02:02.444 Message: lib/net: Defining dependency "net" 00:02:02.444 Message: lib/meter: Defining dependency "meter" 00:02:02.444 Message: lib/ethdev: Defining dependency "ethdev" 00:02:02.444 Message: lib/pci: Defining dependency "pci" 00:02:02.444 Message: lib/cmdline: Defining dependency "cmdline" 00:02:02.444 Message: lib/hash: Defining dependency "hash" 00:02:02.444 Message: lib/timer: Defining dependency "timer" 00:02:02.444 Message: lib/compressdev: Defining dependency "compressdev" 00:02:02.445 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:02.445 Message: lib/dmadev: Defining dependency "dmadev" 00:02:02.445 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:02.445 Message: lib/power: Defining dependency "power" 00:02:02.445 Message: lib/reorder: Defining dependency "reorder" 00:02:02.445 Message: lib/security: Defining dependency "security" 00:02:02.445 Has header "linux/userfaultfd.h" : YES 00:02:02.445 Has header "linux/vduse.h" : YES 00:02:02.445 Message: lib/vhost: Defining dependency "vhost" 00:02:02.445 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:02.445 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:02.445 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:02.445 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:02.445 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:02:02.445 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:02:02.445 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:02:02.445 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:02:02.445 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:02:02.445 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:02:02.445 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:02.445 Configuring doxy-api-html.conf using configuration 00:02:02.445 Configuring doxy-api-man.conf using configuration 00:02:02.445 Program mandb found: YES (/usr/bin/mandb) 00:02:02.445 Program sphinx-build found: NO 00:02:02.445 Configuring rte_build_config.h using configuration 00:02:02.445 Message: 00:02:02.445 ================= 00:02:02.445 Applications Enabled 00:02:02.445 ================= 00:02:02.445 00:02:02.445 apps: 00:02:02.445 00:02:02.445 00:02:02.445 Message: 00:02:02.445 ================= 00:02:02.445 Libraries Enabled 00:02:02.445 ================= 00:02:02.445 00:02:02.445 libs: 00:02:02.445 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:02:02.445 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:02:02.445 cryptodev, dmadev, power, reorder, security, vhost, 00:02:02.445 00:02:02.445 Message: 00:02:02.445 =============== 00:02:02.445 Drivers Enabled 00:02:02.445 =============== 00:02:02.445 00:02:02.445 common: 00:02:02.445 00:02:02.445 bus: 00:02:02.445 pci, vdev, 00:02:02.445 mempool: 00:02:02.445 ring, 00:02:02.445 dma: 00:02:02.445 00:02:02.445 net: 00:02:02.445 00:02:02.445 crypto: 00:02:02.445 00:02:02.445 compress: 00:02:02.445 00:02:02.445 vdpa: 00:02:02.445 00:02:02.445 00:02:02.445 Message: 00:02:02.445 ================= 00:02:02.445 Content Skipped 00:02:02.445 ================= 00:02:02.445 00:02:02.445 apps: 00:02:02.445 dumpcap: explicitly disabled via build config 00:02:02.445 graph: explicitly disabled via build config 00:02:02.445 pdump: explicitly disabled via build config 00:02:02.445 proc-info: explicitly disabled via build config 00:02:02.445 test-acl: explicitly disabled via build config 00:02:02.445 test-bbdev: explicitly disabled via build config 00:02:02.445 test-cmdline: explicitly disabled via build config 00:02:02.445 test-compress-perf: explicitly disabled via build config 00:02:02.445 test-crypto-perf: explicitly disabled via build config 00:02:02.445 test-dma-perf: explicitly disabled via build config 00:02:02.445 test-eventdev: explicitly disabled via build config 00:02:02.445 test-fib: explicitly disabled via build config 00:02:02.445 test-flow-perf: explicitly disabled via build config 00:02:02.445 test-gpudev: explicitly disabled via build config 00:02:02.445 test-mldev: explicitly disabled via build config 00:02:02.445 test-pipeline: explicitly disabled via build config 00:02:02.445 test-pmd: explicitly disabled via build config 00:02:02.445 test-regex: explicitly disabled via build config 00:02:02.445 test-sad: explicitly disabled via build config 00:02:02.445 test-security-perf: explicitly disabled via build config 00:02:02.445 00:02:02.445 libs: 00:02:02.445 argparse: explicitly disabled via build config 00:02:02.445 metrics: explicitly disabled via build config 00:02:02.445 acl: explicitly disabled via build config 00:02:02.445 bbdev: explicitly disabled via build config 00:02:02.445 bitratestats: explicitly disabled via build config 00:02:02.445 bpf: explicitly disabled via build config 00:02:02.445 cfgfile: explicitly disabled via build config 00:02:02.445 distributor: explicitly disabled via build config 00:02:02.445 efd: explicitly disabled via build config 00:02:02.445 eventdev: explicitly disabled via build config 00:02:02.445 dispatcher: explicitly disabled via build config 00:02:02.445 gpudev: explicitly disabled via build config 00:02:02.445 gro: explicitly disabled via build config 00:02:02.445 gso: explicitly disabled via build config 00:02:02.445 ip_frag: explicitly disabled via build config 00:02:02.445 jobstats: explicitly disabled via build config 00:02:02.445 latencystats: explicitly disabled via build config 00:02:02.445 lpm: explicitly disabled via build config 00:02:02.445 member: explicitly disabled via build config 00:02:02.445 pcapng: explicitly disabled via build config 00:02:02.445 rawdev: explicitly disabled via build config 00:02:02.445 regexdev: explicitly disabled via build config 00:02:02.445 mldev: explicitly disabled via build config 00:02:02.445 rib: explicitly disabled via build config 00:02:02.445 sched: explicitly disabled via build config 00:02:02.445 stack: explicitly disabled via build config 00:02:02.445 ipsec: explicitly disabled via build config 00:02:02.445 pdcp: explicitly disabled via build config 00:02:02.445 fib: explicitly disabled via build config 00:02:02.445 port: explicitly disabled via build config 00:02:02.445 pdump: explicitly disabled via build config 00:02:02.445 table: explicitly disabled via build config 00:02:02.445 pipeline: explicitly disabled via build config 00:02:02.445 graph: explicitly disabled via build config 00:02:02.445 node: explicitly disabled via build config 00:02:02.445 00:02:02.445 drivers: 00:02:02.445 common/cpt: not in enabled drivers build config 00:02:02.445 common/dpaax: not in enabled drivers build config 00:02:02.445 common/iavf: not in enabled drivers build config 00:02:02.445 common/idpf: not in enabled drivers build config 00:02:02.445 common/ionic: not in enabled drivers build config 00:02:02.445 common/mvep: not in enabled drivers build config 00:02:02.445 common/octeontx: not in enabled drivers build config 00:02:02.445 bus/auxiliary: not in enabled drivers build config 00:02:02.445 bus/cdx: not in enabled drivers build config 00:02:02.445 bus/dpaa: not in enabled drivers build config 00:02:02.445 bus/fslmc: not in enabled drivers build config 00:02:02.445 bus/ifpga: not in enabled drivers build config 00:02:02.445 bus/platform: not in enabled drivers build config 00:02:02.445 bus/uacce: not in enabled drivers build config 00:02:02.445 bus/vmbus: not in enabled drivers build config 00:02:02.445 common/cnxk: not in enabled drivers build config 00:02:02.445 common/mlx5: not in enabled drivers build config 00:02:02.445 common/nfp: not in enabled drivers build config 00:02:02.445 common/nitrox: not in enabled drivers build config 00:02:02.445 common/qat: not in enabled drivers build config 00:02:02.445 common/sfc_efx: not in enabled drivers build config 00:02:02.445 mempool/bucket: not in enabled drivers build config 00:02:02.445 mempool/cnxk: not in enabled drivers build config 00:02:02.445 mempool/dpaa: not in enabled drivers build config 00:02:02.445 mempool/dpaa2: not in enabled drivers build config 00:02:02.445 mempool/octeontx: not in enabled drivers build config 00:02:02.445 mempool/stack: not in enabled drivers build config 00:02:02.445 dma/cnxk: not in enabled drivers build config 00:02:02.445 dma/dpaa: not in enabled drivers build config 00:02:02.445 dma/dpaa2: not in enabled drivers build config 00:02:02.445 dma/hisilicon: not in enabled drivers build config 00:02:02.445 dma/idxd: not in enabled drivers build config 00:02:02.445 dma/ioat: not in enabled drivers build config 00:02:02.445 dma/skeleton: not in enabled drivers build config 00:02:02.445 net/af_packet: not in enabled drivers build config 00:02:02.445 net/af_xdp: not in enabled drivers build config 00:02:02.445 net/ark: not in enabled drivers build config 00:02:02.445 net/atlantic: not in enabled drivers build config 00:02:02.445 net/avp: not in enabled drivers build config 00:02:02.445 net/axgbe: not in enabled drivers build config 00:02:02.445 net/bnx2x: not in enabled drivers build config 00:02:02.445 net/bnxt: not in enabled drivers build config 00:02:02.445 net/bonding: not in enabled drivers build config 00:02:02.445 net/cnxk: not in enabled drivers build config 00:02:02.445 net/cpfl: not in enabled drivers build config 00:02:02.445 net/cxgbe: not in enabled drivers build config 00:02:02.445 net/dpaa: not in enabled drivers build config 00:02:02.445 net/dpaa2: not in enabled drivers build config 00:02:02.445 net/e1000: not in enabled drivers build config 00:02:02.445 net/ena: not in enabled drivers build config 00:02:02.445 net/enetc: not in enabled drivers build config 00:02:02.445 net/enetfec: not in enabled drivers build config 00:02:02.445 net/enic: not in enabled drivers build config 00:02:02.445 net/failsafe: not in enabled drivers build config 00:02:02.445 net/fm10k: not in enabled drivers build config 00:02:02.445 net/gve: not in enabled drivers build config 00:02:02.445 net/hinic: not in enabled drivers build config 00:02:02.445 net/hns3: not in enabled drivers build config 00:02:02.445 net/i40e: not in enabled drivers build config 00:02:02.445 net/iavf: not in enabled drivers build config 00:02:02.445 net/ice: not in enabled drivers build config 00:02:02.445 net/idpf: not in enabled drivers build config 00:02:02.445 net/igc: not in enabled drivers build config 00:02:02.445 net/ionic: not in enabled drivers build config 00:02:02.445 net/ipn3ke: not in enabled drivers build config 00:02:02.445 net/ixgbe: not in enabled drivers build config 00:02:02.445 net/mana: not in enabled drivers build config 00:02:02.445 net/memif: not in enabled drivers build config 00:02:02.445 net/mlx4: not in enabled drivers build config 00:02:02.445 net/mlx5: not in enabled drivers build config 00:02:02.445 net/mvneta: not in enabled drivers build config 00:02:02.445 net/mvpp2: not in enabled drivers build config 00:02:02.445 net/netvsc: not in enabled drivers build config 00:02:02.445 net/nfb: not in enabled drivers build config 00:02:02.445 net/nfp: not in enabled drivers build config 00:02:02.445 net/ngbe: not in enabled drivers build config 00:02:02.445 net/null: not in enabled drivers build config 00:02:02.445 net/octeontx: not in enabled drivers build config 00:02:02.445 net/octeon_ep: not in enabled drivers build config 00:02:02.445 net/pcap: not in enabled drivers build config 00:02:02.445 net/pfe: not in enabled drivers build config 00:02:02.445 net/qede: not in enabled drivers build config 00:02:02.445 net/ring: not in enabled drivers build config 00:02:02.445 net/sfc: not in enabled drivers build config 00:02:02.445 net/softnic: not in enabled drivers build config 00:02:02.445 net/tap: not in enabled drivers build config 00:02:02.445 net/thunderx: not in enabled drivers build config 00:02:02.445 net/txgbe: not in enabled drivers build config 00:02:02.445 net/vdev_netvsc: not in enabled drivers build config 00:02:02.445 net/vhost: not in enabled drivers build config 00:02:02.445 net/virtio: not in enabled drivers build config 00:02:02.445 net/vmxnet3: not in enabled drivers build config 00:02:02.445 raw/*: missing internal dependency, "rawdev" 00:02:02.445 crypto/armv8: not in enabled drivers build config 00:02:02.445 crypto/bcmfs: not in enabled drivers build config 00:02:02.445 crypto/caam_jr: not in enabled drivers build config 00:02:02.445 crypto/ccp: not in enabled drivers build config 00:02:02.445 crypto/cnxk: not in enabled drivers build config 00:02:02.445 crypto/dpaa_sec: not in enabled drivers build config 00:02:02.445 crypto/dpaa2_sec: not in enabled drivers build config 00:02:02.445 crypto/ipsec_mb: not in enabled drivers build config 00:02:02.445 crypto/mlx5: not in enabled drivers build config 00:02:02.445 crypto/mvsam: not in enabled drivers build config 00:02:02.445 crypto/nitrox: not in enabled drivers build config 00:02:02.445 crypto/null: not in enabled drivers build config 00:02:02.445 crypto/octeontx: not in enabled drivers build config 00:02:02.445 crypto/openssl: not in enabled drivers build config 00:02:02.445 crypto/scheduler: not in enabled drivers build config 00:02:02.445 crypto/uadk: not in enabled drivers build config 00:02:02.445 crypto/virtio: not in enabled drivers build config 00:02:02.445 compress/isal: not in enabled drivers build config 00:02:02.445 compress/mlx5: not in enabled drivers build config 00:02:02.445 compress/nitrox: not in enabled drivers build config 00:02:02.445 compress/octeontx: not in enabled drivers build config 00:02:02.445 compress/zlib: not in enabled drivers build config 00:02:02.445 regex/*: missing internal dependency, "regexdev" 00:02:02.445 ml/*: missing internal dependency, "mldev" 00:02:02.445 vdpa/ifc: not in enabled drivers build config 00:02:02.445 vdpa/mlx5: not in enabled drivers build config 00:02:02.445 vdpa/nfp: not in enabled drivers build config 00:02:02.445 vdpa/sfc: not in enabled drivers build config 00:02:02.445 event/*: missing internal dependency, "eventdev" 00:02:02.445 baseband/*: missing internal dependency, "bbdev" 00:02:02.445 gpu/*: missing internal dependency, "gpudev" 00:02:02.445 00:02:02.445 00:02:02.445 Build targets in project: 84 00:02:02.445 00:02:02.445 DPDK 24.03.0 00:02:02.445 00:02:02.445 User defined options 00:02:02.445 buildtype : debug 00:02:02.445 default_library : shared 00:02:02.445 libdir : lib 00:02:02.445 prefix : /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:02:02.445 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:02:02.445 c_link_args : 00:02:02.445 cpu_instruction_set: native 00:02:02.445 disable_apps : test-dma-perf,test,test-sad,test-acl,test-pmd,test-mldev,test-compress-perf,test-cmdline,test-regex,test-fib,graph,test-bbdev,dumpcap,test-gpudev,proc-info,test-pipeline,test-flow-perf,test-crypto-perf,pdump,test-eventdev,test-security-perf 00:02:02.445 disable_libs : port,lpm,ipsec,regexdev,dispatcher,argparse,bitratestats,rawdev,stack,graph,acl,bbdev,pipeline,member,sched,pcapng,mldev,eventdev,efd,metrics,latencystats,cfgfile,ip_frag,jobstats,pdump,pdcp,rib,node,fib,distributor,gso,table,bpf,gpudev,gro 00:02:02.445 enable_docs : false 00:02:02.445 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm 00:02:02.445 enable_kmods : false 00:02:02.445 max_lcores : 128 00:02:02.445 tests : false 00:02:02.445 00:02:02.445 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:02.445 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp' 00:02:02.445 [1/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:02.445 [2/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:02.445 [3/267] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:02:02.445 [4/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:02.445 [5/267] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:02.445 [6/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:02.445 [7/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:02.445 [8/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:02.445 [9/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:02.445 [10/267] Linking static target lib/librte_kvargs.a 00:02:02.445 [11/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:02.445 [12/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:02.445 [13/267] Compiling C object lib/librte_log.a.p/log_log.c.o 00:02:02.445 [14/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:02.445 [15/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:02.705 [16/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:02.705 [17/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:02.705 [18/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:02.705 [19/267] Linking static target lib/librte_log.a 00:02:02.705 [20/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:02.705 [21/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:02.705 [22/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:02.705 [23/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:02.705 [24/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:02.705 [25/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:02.705 [26/267] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:02.705 [27/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:02.705 [28/267] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:02:02.705 [29/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:02.705 [30/267] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:02.705 [31/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:02.705 [32/267] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:02.705 [33/267] Linking static target lib/librte_pci.a 00:02:02.705 [34/267] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:02.705 [35/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:02.705 [36/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:02.705 [37/267] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:02.705 [38/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:02.963 [39/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:02.963 [40/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:02.963 [41/267] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:02.963 [42/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:02.963 [43/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:02.963 [44/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:02.963 [45/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:02.963 [46/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:02.963 [47/267] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:02.963 [48/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:02.963 [49/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:02.963 [50/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:02.963 [51/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:02.963 [52/267] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:02.963 [53/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:02.963 [54/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:02.963 [55/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:02.963 [56/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:02.963 [57/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:02.963 [58/267] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:02.963 [59/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:02.963 [60/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:02.963 [61/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:02.963 [62/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:02.963 [63/267] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:02.963 [64/267] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:02.963 [65/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:02:02.963 [66/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:02.963 [67/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:02.963 [68/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:02.963 [69/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:02.963 [70/267] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:02.963 [71/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:02.963 [72/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:02.963 [73/267] Compiling C object lib/librte_net.a.p/net_net_crc_avx512.c.o 00:02:02.963 [74/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:02.963 [75/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:02.963 [76/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:02.963 [77/267] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:02.963 [78/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:02.963 [79/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:02.963 [80/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:02.963 [81/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:02.963 [82/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:02.963 [83/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:02.963 [84/267] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:02.963 [85/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:02.963 [86/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:02.963 [87/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:02.963 [88/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:02.963 [89/267] Linking static target lib/librte_telemetry.a 00:02:02.963 [90/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:02.963 [91/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:02.963 [92/267] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:02.963 [93/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:02.963 [94/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:03.222 [95/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:02:03.222 [96/267] Linking static target lib/librte_meter.a 00:02:03.222 [97/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:03.222 [98/267] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:03.222 [99/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:03.222 [100/267] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:03.222 [101/267] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:03.222 [102/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:03.222 [103/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:03.222 [104/267] Linking static target lib/librte_timer.a 00:02:03.222 [105/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:03.222 [106/267] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:03.222 [107/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:02:03.222 [108/267] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:02:03.222 [109/267] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:02:03.222 [110/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:03.222 [111/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:03.222 [112/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:03.222 [113/267] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:03.222 [114/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:03.222 [115/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:03.222 [116/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:03.222 [117/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:03.222 [118/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:03.222 [119/267] Linking static target lib/librte_ring.a 00:02:03.222 [120/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:03.222 [121/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:03.222 [122/267] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:03.223 [123/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:03.223 [124/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:03.223 [125/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:03.223 [126/267] Linking static target lib/librte_cmdline.a 00:02:03.223 [127/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:02:03.223 [128/267] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:02:03.223 [129/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:03.223 [130/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:03.223 [131/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:03.223 [132/267] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:03.223 [133/267] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:03.223 [134/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:03.223 [135/267] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:03.223 [136/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:03.223 [137/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:03.223 [138/267] Linking static target lib/librte_power.a 00:02:03.223 [139/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:03.223 [140/267] Linking static target lib/librte_rcu.a 00:02:03.223 [141/267] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:03.223 [142/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:03.223 [143/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:03.223 [144/267] Linking static target lib/librte_net.a 00:02:03.223 [145/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:03.223 [146/267] Linking static target lib/librte_mempool.a 00:02:03.223 [147/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:03.223 [148/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:03.223 [149/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:03.223 [150/267] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.223 [151/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:03.223 [152/267] Linking static target lib/librte_dmadev.a 00:02:03.223 [153/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:03.223 [154/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:03.223 [155/267] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:03.223 [156/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:03.223 [157/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:03.223 [158/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:02:03.223 [159/267] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:03.223 [160/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:03.223 [161/267] Linking static target lib/librte_compressdev.a 00:02:03.223 [162/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:03.223 [163/267] Linking target lib/librte_log.so.24.1 00:02:03.223 [164/267] Linking static target lib/librte_security.a 00:02:03.223 [165/267] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:03.223 [166/267] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:03.223 [167/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:03.223 [168/267] Linking static target lib/librte_eal.a 00:02:03.223 [169/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:03.223 [170/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:03.223 [171/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:03.223 [172/267] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:03.223 [173/267] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:03.223 [174/267] Linking static target lib/librte_mbuf.a 00:02:03.223 [175/267] Linking static target lib/librte_reorder.a 00:02:03.223 [176/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:03.223 [177/267] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:03.223 [178/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:03.223 [179/267] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:03.483 [180/267] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:02:03.483 [181/267] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:03.483 [182/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:03.483 [183/267] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:02:03.483 [184/267] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.483 [185/267] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:03.483 [186/267] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:03.483 [187/267] Linking static target lib/librte_hash.a 00:02:03.483 [188/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:03.483 [189/267] Linking target lib/librte_kvargs.so.24.1 00:02:03.483 [190/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:03.483 [191/267] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:03.483 [192/267] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:03.483 [193/267] Linking static target drivers/librte_bus_vdev.a 00:02:03.483 [194/267] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:03.483 [195/267] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:03.483 [196/267] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.483 [197/267] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:03.483 [198/267] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:03.483 [199/267] Linking static target drivers/librte_mempool_ring.a 00:02:03.483 [200/267] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:03.483 [201/267] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:03.483 [202/267] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.483 [203/267] Linking static target drivers/librte_bus_pci.a 00:02:03.483 [204/267] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:02:03.483 [205/267] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.483 [206/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:03.744 [207/267] Linking static target lib/librte_cryptodev.a 00:02:03.744 [208/267] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.744 [209/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:03.744 [210/267] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.744 [211/267] Linking target lib/librte_telemetry.so.24.1 00:02:03.744 [212/267] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.744 [213/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:04.006 [214/267] Linking static target lib/librte_ethdev.a 00:02:04.006 [215/267] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.006 [216/267] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:02:04.006 [217/267] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.006 [218/267] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.006 [219/267] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.006 [220/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:04.267 [221/267] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.268 [222/267] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.268 [223/267] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.529 [224/267] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.529 [225/267] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.529 [226/267] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:05.101 [227/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:02:05.101 [228/267] Linking static target lib/librte_vhost.a 00:02:06.045 [229/267] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:06.985 [230/267] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:02:13.560 [231/267] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:14.940 [232/267] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:14.941 [233/267] Linking target lib/librte_eal.so.24.1 00:02:14.941 [234/267] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:02:14.941 [235/267] Linking target lib/librte_meter.so.24.1 00:02:14.941 [236/267] Linking target lib/librte_ring.so.24.1 00:02:14.941 [237/267] Linking target lib/librte_pci.so.24.1 00:02:14.941 [238/267] Linking target lib/librte_dmadev.so.24.1 00:02:14.941 [239/267] Linking target lib/librte_timer.so.24.1 00:02:14.941 [240/267] Linking target drivers/librte_bus_vdev.so.24.1 00:02:15.201 [241/267] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:02:15.201 [242/267] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:02:15.201 [243/267] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:02:15.201 [244/267] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:02:15.201 [245/267] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:02:15.201 [246/267] Linking target drivers/librte_bus_pci.so.24.1 00:02:15.201 [247/267] Linking target lib/librte_rcu.so.24.1 00:02:15.201 [248/267] Linking target lib/librte_mempool.so.24.1 00:02:15.201 [249/267] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:02:15.201 [250/267] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:02:15.461 [251/267] Linking target drivers/librte_mempool_ring.so.24.1 00:02:15.461 [252/267] Linking target lib/librte_mbuf.so.24.1 00:02:15.461 [253/267] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:02:15.461 [254/267] Linking target lib/librte_net.so.24.1 00:02:15.461 [255/267] Linking target lib/librte_compressdev.so.24.1 00:02:15.461 [256/267] Linking target lib/librte_reorder.so.24.1 00:02:15.461 [257/267] Linking target lib/librte_cryptodev.so.24.1 00:02:15.721 [258/267] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:02:15.721 [259/267] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:02:15.721 [260/267] Linking target lib/librte_hash.so.24.1 00:02:15.721 [261/267] Linking target lib/librte_cmdline.so.24.1 00:02:15.721 [262/267] Linking target lib/librte_security.so.24.1 00:02:15.721 [263/267] Linking target lib/librte_ethdev.so.24.1 00:02:15.721 [264/267] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:02:15.981 [265/267] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:02:15.981 [266/267] Linking target lib/librte_power.so.24.1 00:02:15.981 [267/267] Linking target lib/librte_vhost.so.24.1 00:02:15.981 INFO: autodetecting backend as ninja 00:02:15.981 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp -j 144 00:02:20.182 CC lib/ut_mock/mock.o 00:02:20.182 CC lib/ut/ut.o 00:02:20.182 CC lib/log/log.o 00:02:20.182 CC lib/log/log_flags.o 00:02:20.182 CC lib/log/log_deprecated.o 00:02:20.182 LIB libspdk_ut.a 00:02:20.182 LIB libspdk_ut_mock.a 00:02:20.182 LIB libspdk_log.a 00:02:20.182 SO libspdk_ut.so.2.0 00:02:20.183 SO libspdk_ut_mock.so.6.0 00:02:20.183 SO libspdk_log.so.7.1 00:02:20.183 SYMLINK libspdk_ut.so 00:02:20.183 SYMLINK libspdk_ut_mock.so 00:02:20.183 SYMLINK libspdk_log.so 00:02:20.183 CC lib/util/base64.o 00:02:20.183 CC lib/util/bit_array.o 00:02:20.183 CC lib/util/cpuset.o 00:02:20.183 CC lib/util/crc32.o 00:02:20.183 CC lib/util/crc16.o 00:02:20.183 CC lib/util/crc32c.o 00:02:20.183 CC lib/util/crc32_ieee.o 00:02:20.183 CXX lib/trace_parser/trace.o 00:02:20.183 CC lib/util/crc64.o 00:02:20.183 CC lib/util/dif.o 00:02:20.183 CC lib/util/fd.o 00:02:20.183 CC lib/dma/dma.o 00:02:20.183 CC lib/util/fd_group.o 00:02:20.183 CC lib/ioat/ioat.o 00:02:20.183 CC lib/util/file.o 00:02:20.183 CC lib/util/hexlify.o 00:02:20.183 CC lib/util/iov.o 00:02:20.183 CC lib/util/math.o 00:02:20.183 CC lib/util/net.o 00:02:20.183 CC lib/util/pipe.o 00:02:20.183 CC lib/util/strerror_tls.o 00:02:20.183 CC lib/util/string.o 00:02:20.183 CC lib/util/uuid.o 00:02:20.183 CC lib/util/xor.o 00:02:20.183 CC lib/util/zipf.o 00:02:20.183 CC lib/util/md5.o 00:02:20.444 CC lib/vfio_user/host/vfio_user_pci.o 00:02:20.444 CC lib/vfio_user/host/vfio_user.o 00:02:20.444 LIB libspdk_dma.a 00:02:20.444 SO libspdk_dma.so.5.0 00:02:20.444 LIB libspdk_ioat.a 00:02:20.444 SO libspdk_ioat.so.7.0 00:02:20.444 SYMLINK libspdk_dma.so 00:02:20.444 SYMLINK libspdk_ioat.so 00:02:20.444 LIB libspdk_vfio_user.a 00:02:20.705 SO libspdk_vfio_user.so.5.0 00:02:20.705 SYMLINK libspdk_vfio_user.so 00:02:20.705 LIB libspdk_util.a 00:02:20.705 SO libspdk_util.so.10.1 00:02:20.967 SYMLINK libspdk_util.so 00:02:20.967 LIB libspdk_trace_parser.a 00:02:20.967 SO libspdk_trace_parser.so.6.0 00:02:20.967 SYMLINK libspdk_trace_parser.so 00:02:21.227 CC lib/vmd/vmd.o 00:02:21.227 CC lib/vmd/led.o 00:02:21.227 CC lib/env_dpdk/env.o 00:02:21.227 CC lib/env_dpdk/memory.o 00:02:21.227 CC lib/env_dpdk/threads.o 00:02:21.227 CC lib/env_dpdk/pci.o 00:02:21.227 CC lib/env_dpdk/init.o 00:02:21.227 CC lib/env_dpdk/pci_ioat.o 00:02:21.227 CC lib/env_dpdk/pci_virtio.o 00:02:21.227 CC lib/env_dpdk/pci_vmd.o 00:02:21.227 CC lib/rdma_utils/rdma_utils.o 00:02:21.227 CC lib/json/json_parse.o 00:02:21.227 CC lib/env_dpdk/pci_idxd.o 00:02:21.227 CC lib/json/json_write.o 00:02:21.227 CC lib/conf/conf.o 00:02:21.227 CC lib/env_dpdk/pci_event.o 00:02:21.227 CC lib/json/json_util.o 00:02:21.228 CC lib/env_dpdk/sigbus_handler.o 00:02:21.228 CC lib/idxd/idxd.o 00:02:21.228 CC lib/idxd/idxd_kernel.o 00:02:21.228 CC lib/env_dpdk/pci_dpdk.o 00:02:21.228 CC lib/env_dpdk/pci_dpdk_2207.o 00:02:21.228 CC lib/idxd/idxd_user.o 00:02:21.228 CC lib/env_dpdk/pci_dpdk_2211.o 00:02:21.489 LIB libspdk_conf.a 00:02:21.489 SO libspdk_conf.so.6.0 00:02:21.489 LIB libspdk_json.a 00:02:21.489 LIB libspdk_rdma_utils.a 00:02:21.489 SYMLINK libspdk_conf.so 00:02:21.489 SO libspdk_json.so.6.0 00:02:21.489 SO libspdk_rdma_utils.so.1.0 00:02:21.489 SYMLINK libspdk_rdma_utils.so 00:02:21.489 SYMLINK libspdk_json.so 00:02:21.751 LIB libspdk_idxd.a 00:02:21.751 LIB libspdk_vmd.a 00:02:21.751 SO libspdk_idxd.so.12.1 00:02:21.751 SO libspdk_vmd.so.6.0 00:02:22.013 SYMLINK libspdk_idxd.so 00:02:22.013 SYMLINK libspdk_vmd.so 00:02:22.013 CC lib/jsonrpc/jsonrpc_server.o 00:02:22.013 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:02:22.013 CC lib/jsonrpc/jsonrpc_client.o 00:02:22.013 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:02:22.013 CC lib/rdma_provider/common.o 00:02:22.013 CC lib/rdma_provider/rdma_provider_verbs.o 00:02:22.274 LIB libspdk_rdma_provider.a 00:02:22.274 LIB libspdk_jsonrpc.a 00:02:22.274 SO libspdk_rdma_provider.so.7.0 00:02:22.274 SO libspdk_jsonrpc.so.6.0 00:02:22.274 SYMLINK libspdk_rdma_provider.so 00:02:22.274 SYMLINK libspdk_jsonrpc.so 00:02:22.535 LIB libspdk_env_dpdk.a 00:02:22.535 SO libspdk_env_dpdk.so.15.1 00:02:22.535 CC lib/rpc/rpc.o 00:02:22.535 SYMLINK libspdk_env_dpdk.so 00:02:22.796 LIB libspdk_rpc.a 00:02:22.796 SO libspdk_rpc.so.6.0 00:02:23.058 SYMLINK libspdk_rpc.so 00:02:23.320 CC lib/keyring/keyring.o 00:02:23.320 CC lib/keyring/keyring_rpc.o 00:02:23.320 CC lib/notify/notify.o 00:02:23.320 CC lib/notify/notify_rpc.o 00:02:23.320 CC lib/trace/trace.o 00:02:23.320 CC lib/trace/trace_flags.o 00:02:23.320 CC lib/trace/trace_rpc.o 00:02:23.581 LIB libspdk_notify.a 00:02:23.581 SO libspdk_notify.so.6.0 00:02:23.581 LIB libspdk_keyring.a 00:02:23.581 LIB libspdk_trace.a 00:02:23.581 SO libspdk_keyring.so.2.0 00:02:23.581 SYMLINK libspdk_notify.so 00:02:23.581 SO libspdk_trace.so.11.0 00:02:23.581 SYMLINK libspdk_keyring.so 00:02:23.581 SYMLINK libspdk_trace.so 00:02:24.154 CC lib/thread/thread.o 00:02:24.154 CC lib/thread/iobuf.o 00:02:24.154 CC lib/sock/sock.o 00:02:24.154 CC lib/sock/sock_rpc.o 00:02:24.415 LIB libspdk_sock.a 00:02:24.415 SO libspdk_sock.so.10.0 00:02:24.415 SYMLINK libspdk_sock.so 00:02:24.996 CC lib/nvme/nvme_ctrlr_cmd.o 00:02:24.996 CC lib/nvme/nvme_ctrlr.o 00:02:24.996 CC lib/nvme/nvme_fabric.o 00:02:24.996 CC lib/nvme/nvme_ns_cmd.o 00:02:24.996 CC lib/nvme/nvme_ns.o 00:02:24.996 CC lib/nvme/nvme_pcie_common.o 00:02:24.996 CC lib/nvme/nvme_pcie.o 00:02:24.996 CC lib/nvme/nvme_qpair.o 00:02:24.996 CC lib/nvme/nvme.o 00:02:24.996 CC lib/nvme/nvme_quirks.o 00:02:24.996 CC lib/nvme/nvme_transport.o 00:02:24.996 CC lib/nvme/nvme_discovery.o 00:02:24.997 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:02:24.997 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:02:24.997 CC lib/nvme/nvme_tcp.o 00:02:24.997 CC lib/nvme/nvme_opal.o 00:02:24.997 CC lib/nvme/nvme_io_msg.o 00:02:24.997 CC lib/nvme/nvme_poll_group.o 00:02:24.997 CC lib/nvme/nvme_zns.o 00:02:24.997 CC lib/nvme/nvme_stubs.o 00:02:24.997 CC lib/nvme/nvme_auth.o 00:02:24.997 CC lib/nvme/nvme_cuse.o 00:02:24.997 CC lib/nvme/nvme_vfio_user.o 00:02:24.997 CC lib/nvme/nvme_rdma.o 00:02:25.259 LIB libspdk_thread.a 00:02:25.259 SO libspdk_thread.so.11.0 00:02:25.522 SYMLINK libspdk_thread.so 00:02:25.783 CC lib/blob/blobstore.o 00:02:25.783 CC lib/virtio/virtio.o 00:02:25.783 CC lib/virtio/virtio_vhost_user.o 00:02:25.783 CC lib/virtio/virtio_vfio_user.o 00:02:25.783 CC lib/blob/request.o 00:02:25.783 CC lib/virtio/virtio_pci.o 00:02:25.783 CC lib/blob/zeroes.o 00:02:25.783 CC lib/blob/blob_bs_dev.o 00:02:25.783 CC lib/accel/accel.o 00:02:25.783 CC lib/vfu_tgt/tgt_endpoint.o 00:02:25.783 CC lib/vfu_tgt/tgt_rpc.o 00:02:25.783 CC lib/accel/accel_rpc.o 00:02:25.783 CC lib/accel/accel_sw.o 00:02:25.783 CC lib/init/json_config.o 00:02:25.783 CC lib/init/subsystem.o 00:02:25.783 CC lib/init/subsystem_rpc.o 00:02:25.783 CC lib/init/rpc.o 00:02:25.783 CC lib/fsdev/fsdev.o 00:02:25.783 CC lib/fsdev/fsdev_io.o 00:02:25.783 CC lib/fsdev/fsdev_rpc.o 00:02:26.043 LIB libspdk_init.a 00:02:26.043 SO libspdk_init.so.6.0 00:02:26.043 LIB libspdk_virtio.a 00:02:26.043 LIB libspdk_vfu_tgt.a 00:02:26.043 SYMLINK libspdk_init.so 00:02:26.043 SO libspdk_virtio.so.7.0 00:02:26.043 SO libspdk_vfu_tgt.so.3.0 00:02:26.305 SYMLINK libspdk_virtio.so 00:02:26.305 SYMLINK libspdk_vfu_tgt.so 00:02:26.305 LIB libspdk_fsdev.a 00:02:26.567 SO libspdk_fsdev.so.2.0 00:02:26.567 CC lib/event/app.o 00:02:26.567 CC lib/event/reactor.o 00:02:26.567 CC lib/event/log_rpc.o 00:02:26.567 CC lib/event/app_rpc.o 00:02:26.567 CC lib/event/scheduler_static.o 00:02:26.567 SYMLINK libspdk_fsdev.so 00:02:26.828 LIB libspdk_accel.a 00:02:26.828 SO libspdk_accel.so.16.0 00:02:26.828 LIB libspdk_nvme.a 00:02:26.828 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:02:26.828 SYMLINK libspdk_accel.so 00:02:26.828 LIB libspdk_event.a 00:02:26.828 SO libspdk_event.so.14.0 00:02:27.089 SO libspdk_nvme.so.15.0 00:02:27.089 SYMLINK libspdk_event.so 00:02:27.089 SYMLINK libspdk_nvme.so 00:02:27.409 CC lib/bdev/bdev.o 00:02:27.409 CC lib/bdev/bdev_rpc.o 00:02:27.409 CC lib/bdev/part.o 00:02:27.409 CC lib/bdev/bdev_zone.o 00:02:27.409 CC lib/bdev/scsi_nvme.o 00:02:27.409 LIB libspdk_fuse_dispatcher.a 00:02:27.409 SO libspdk_fuse_dispatcher.so.1.0 00:02:27.736 SYMLINK libspdk_fuse_dispatcher.so 00:02:28.309 LIB libspdk_blob.a 00:02:28.572 SO libspdk_blob.so.12.0 00:02:28.572 SYMLINK libspdk_blob.so 00:02:28.833 CC lib/lvol/lvol.o 00:02:28.833 CC lib/blobfs/blobfs.o 00:02:28.833 CC lib/blobfs/tree.o 00:02:29.778 LIB libspdk_bdev.a 00:02:29.778 LIB libspdk_blobfs.a 00:02:29.778 SO libspdk_bdev.so.17.0 00:02:29.778 SO libspdk_blobfs.so.11.0 00:02:29.778 LIB libspdk_lvol.a 00:02:29.778 SO libspdk_lvol.so.11.0 00:02:29.778 SYMLINK libspdk_blobfs.so 00:02:29.778 SYMLINK libspdk_bdev.so 00:02:29.778 SYMLINK libspdk_lvol.so 00:02:30.038 CC lib/nvmf/ctrlr_discovery.o 00:02:30.038 CC lib/nvmf/ctrlr.o 00:02:30.038 CC lib/nvmf/ctrlr_bdev.o 00:02:30.038 CC lib/nvmf/subsystem.o 00:02:30.038 CC lib/nvmf/nvmf_rpc.o 00:02:30.038 CC lib/nvmf/nvmf.o 00:02:30.038 CC lib/nvmf/transport.o 00:02:30.038 CC lib/nvmf/stubs.o 00:02:30.038 CC lib/nvmf/tcp.o 00:02:30.038 CC lib/nvmf/mdns_server.o 00:02:30.038 CC lib/nvmf/vfio_user.o 00:02:30.038 CC lib/nvmf/rdma.o 00:02:30.038 CC lib/nvmf/auth.o 00:02:30.038 CC lib/nbd/nbd.o 00:02:30.038 CC lib/nbd/nbd_rpc.o 00:02:30.038 CC lib/scsi/dev.o 00:02:30.038 CC lib/scsi/lun.o 00:02:30.038 CC lib/ublk/ublk.o 00:02:30.038 CC lib/ublk/ublk_rpc.o 00:02:30.038 CC lib/scsi/port.o 00:02:30.038 CC lib/scsi/scsi.o 00:02:30.038 CC lib/scsi/scsi_bdev.o 00:02:30.038 CC lib/scsi/scsi_pr.o 00:02:30.038 CC lib/scsi/scsi_rpc.o 00:02:30.038 CC lib/scsi/task.o 00:02:30.039 CC lib/ftl/ftl_core.o 00:02:30.039 CC lib/ftl/ftl_init.o 00:02:30.039 CC lib/ftl/ftl_layout.o 00:02:30.039 CC lib/ftl/ftl_debug.o 00:02:30.039 CC lib/ftl/ftl_io.o 00:02:30.039 CC lib/ftl/ftl_sb.o 00:02:30.298 CC lib/ftl/ftl_l2p.o 00:02:30.298 CC lib/ftl/ftl_l2p_flat.o 00:02:30.298 CC lib/ftl/ftl_nv_cache.o 00:02:30.298 CC lib/ftl/ftl_band.o 00:02:30.298 CC lib/ftl/ftl_band_ops.o 00:02:30.298 CC lib/ftl/ftl_writer.o 00:02:30.298 CC lib/ftl/ftl_rq.o 00:02:30.298 CC lib/ftl/ftl_reloc.o 00:02:30.298 CC lib/ftl/ftl_l2p_cache.o 00:02:30.298 CC lib/ftl/ftl_p2l.o 00:02:30.298 CC lib/ftl/ftl_p2l_log.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_startup.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_md.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_misc.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_band.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:02:30.298 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:02:30.298 CC lib/ftl/utils/ftl_md.o 00:02:30.298 CC lib/ftl/utils/ftl_conf.o 00:02:30.298 CC lib/ftl/utils/ftl_bitmap.o 00:02:30.298 CC lib/ftl/utils/ftl_mempool.o 00:02:30.298 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:02:30.298 CC lib/ftl/utils/ftl_property.o 00:02:30.298 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:02:30.298 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:02:30.298 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:02:30.298 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:02:30.298 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:02:30.298 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:02:30.298 CC lib/ftl/upgrade/ftl_sb_v3.o 00:02:30.298 CC lib/ftl/upgrade/ftl_sb_v5.o 00:02:30.298 CC lib/ftl/nvc/ftl_nvc_dev.o 00:02:30.298 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:02:30.298 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:02:30.298 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:02:30.298 CC lib/ftl/base/ftl_base_bdev.o 00:02:30.298 CC lib/ftl/base/ftl_base_dev.o 00:02:30.298 CC lib/ftl/ftl_trace.o 00:02:30.558 LIB libspdk_nbd.a 00:02:30.818 SO libspdk_nbd.so.7.0 00:02:30.818 LIB libspdk_scsi.a 00:02:30.818 SYMLINK libspdk_nbd.so 00:02:30.818 SO libspdk_scsi.so.9.0 00:02:30.818 LIB libspdk_ublk.a 00:02:30.818 SYMLINK libspdk_scsi.so 00:02:30.818 SO libspdk_ublk.so.3.0 00:02:31.080 SYMLINK libspdk_ublk.so 00:02:31.080 LIB libspdk_ftl.a 00:02:31.341 CC lib/iscsi/conn.o 00:02:31.341 CC lib/vhost/vhost.o 00:02:31.341 CC lib/iscsi/init_grp.o 00:02:31.341 CC lib/vhost/vhost_rpc.o 00:02:31.341 CC lib/iscsi/iscsi.o 00:02:31.341 CC lib/vhost/vhost_scsi.o 00:02:31.341 CC lib/iscsi/param.o 00:02:31.341 CC lib/vhost/vhost_blk.o 00:02:31.341 CC lib/iscsi/portal_grp.o 00:02:31.341 CC lib/iscsi/tgt_node.o 00:02:31.341 CC lib/vhost/rte_vhost_user.o 00:02:31.341 CC lib/iscsi/iscsi_subsystem.o 00:02:31.341 CC lib/iscsi/iscsi_rpc.o 00:02:31.341 CC lib/iscsi/task.o 00:02:31.341 SO libspdk_ftl.so.9.0 00:02:31.602 SYMLINK libspdk_ftl.so 00:02:31.864 LIB libspdk_nvmf.a 00:02:32.127 SO libspdk_nvmf.so.20.0 00:02:32.127 LIB libspdk_vhost.a 00:02:32.127 SO libspdk_vhost.so.8.0 00:02:32.127 SYMLINK libspdk_nvmf.so 00:02:32.388 SYMLINK libspdk_vhost.so 00:02:32.388 LIB libspdk_iscsi.a 00:02:32.388 SO libspdk_iscsi.so.8.0 00:02:32.652 SYMLINK libspdk_iscsi.so 00:02:33.225 CC module/env_dpdk/env_dpdk_rpc.o 00:02:33.225 CC module/vfu_device/vfu_virtio.o 00:02:33.225 CC module/vfu_device/vfu_virtio_blk.o 00:02:33.225 CC module/vfu_device/vfu_virtio_scsi.o 00:02:33.225 CC module/vfu_device/vfu_virtio_rpc.o 00:02:33.225 CC module/vfu_device/vfu_virtio_fs.o 00:02:33.225 LIB libspdk_env_dpdk_rpc.a 00:02:33.225 CC module/blob/bdev/blob_bdev.o 00:02:33.225 CC module/keyring/linux/keyring_rpc.o 00:02:33.225 CC module/keyring/file/keyring.o 00:02:33.225 CC module/keyring/file/keyring_rpc.o 00:02:33.225 CC module/keyring/linux/keyring.o 00:02:33.225 CC module/scheduler/dynamic/scheduler_dynamic.o 00:02:33.225 CC module/sock/posix/posix.o 00:02:33.225 CC module/accel/error/accel_error.o 00:02:33.225 CC module/scheduler/gscheduler/gscheduler.o 00:02:33.225 CC module/accel/error/accel_error_rpc.o 00:02:33.225 CC module/accel/iaa/accel_iaa_rpc.o 00:02:33.225 CC module/accel/iaa/accel_iaa.o 00:02:33.225 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:02:33.225 CC module/fsdev/aio/fsdev_aio.o 00:02:33.225 CC module/fsdev/aio/fsdev_aio_rpc.o 00:02:33.226 CC module/accel/ioat/accel_ioat.o 00:02:33.226 CC module/fsdev/aio/linux_aio_mgr.o 00:02:33.226 CC module/accel/ioat/accel_ioat_rpc.o 00:02:33.226 CC module/accel/dsa/accel_dsa.o 00:02:33.226 CC module/accel/dsa/accel_dsa_rpc.o 00:02:33.226 SO libspdk_env_dpdk_rpc.so.6.0 00:02:33.487 SYMLINK libspdk_env_dpdk_rpc.so 00:02:33.487 LIB libspdk_keyring_file.a 00:02:33.487 LIB libspdk_keyring_linux.a 00:02:33.487 LIB libspdk_scheduler_gscheduler.a 00:02:33.487 SO libspdk_keyring_file.so.2.0 00:02:33.487 SO libspdk_keyring_linux.so.1.0 00:02:33.487 SO libspdk_scheduler_gscheduler.so.4.0 00:02:33.487 LIB libspdk_scheduler_dpdk_governor.a 00:02:33.487 LIB libspdk_scheduler_dynamic.a 00:02:33.487 LIB libspdk_accel_error.a 00:02:33.487 LIB libspdk_accel_ioat.a 00:02:33.487 SO libspdk_accel_error.so.2.0 00:02:33.487 LIB libspdk_accel_iaa.a 00:02:33.487 SO libspdk_scheduler_dpdk_governor.so.4.0 00:02:33.487 SO libspdk_scheduler_dynamic.so.4.0 00:02:33.487 SYMLINK libspdk_keyring_file.so 00:02:33.487 SYMLINK libspdk_keyring_linux.so 00:02:33.487 SO libspdk_accel_ioat.so.6.0 00:02:33.487 SYMLINK libspdk_scheduler_gscheduler.so 00:02:33.749 LIB libspdk_blob_bdev.a 00:02:33.749 SO libspdk_accel_iaa.so.3.0 00:02:33.749 SYMLINK libspdk_scheduler_dynamic.so 00:02:33.749 SYMLINK libspdk_scheduler_dpdk_governor.so 00:02:33.749 LIB libspdk_accel_dsa.a 00:02:33.749 SO libspdk_blob_bdev.so.12.0 00:02:33.749 SYMLINK libspdk_accel_error.so 00:02:33.749 SYMLINK libspdk_accel_ioat.so 00:02:33.749 SO libspdk_accel_dsa.so.5.0 00:02:33.749 SYMLINK libspdk_accel_iaa.so 00:02:33.749 SYMLINK libspdk_blob_bdev.so 00:02:33.749 LIB libspdk_vfu_device.a 00:02:33.749 SYMLINK libspdk_accel_dsa.so 00:02:33.749 SO libspdk_vfu_device.so.3.0 00:02:33.749 SYMLINK libspdk_vfu_device.so 00:02:34.011 LIB libspdk_fsdev_aio.a 00:02:34.011 SO libspdk_fsdev_aio.so.1.0 00:02:34.011 LIB libspdk_sock_posix.a 00:02:34.011 SO libspdk_sock_posix.so.6.0 00:02:34.011 SYMLINK libspdk_fsdev_aio.so 00:02:34.272 SYMLINK libspdk_sock_posix.so 00:02:34.272 CC module/bdev/malloc/bdev_malloc_rpc.o 00:02:34.272 CC module/bdev/malloc/bdev_malloc.o 00:02:34.272 CC module/bdev/passthru/vbdev_passthru.o 00:02:34.272 CC module/blobfs/bdev/blobfs_bdev.o 00:02:34.272 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:02:34.272 CC module/bdev/ftl/bdev_ftl.o 00:02:34.272 CC module/bdev/ftl/bdev_ftl_rpc.o 00:02:34.272 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:02:34.272 CC module/bdev/nvme/bdev_nvme.o 00:02:34.272 CC module/bdev/error/vbdev_error.o 00:02:34.272 CC module/bdev/error/vbdev_error_rpc.o 00:02:34.272 CC module/bdev/nvme/bdev_nvme_rpc.o 00:02:34.272 CC module/bdev/nvme/nvme_rpc.o 00:02:34.272 CC module/bdev/nvme/bdev_mdns_client.o 00:02:34.272 CC module/bdev/zone_block/vbdev_zone_block.o 00:02:34.272 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:02:34.272 CC module/bdev/delay/vbdev_delay.o 00:02:34.272 CC module/bdev/nvme/vbdev_opal.o 00:02:34.272 CC module/bdev/delay/vbdev_delay_rpc.o 00:02:34.272 CC module/bdev/nvme/vbdev_opal_rpc.o 00:02:34.272 CC module/bdev/split/vbdev_split.o 00:02:34.272 CC module/bdev/split/vbdev_split_rpc.o 00:02:34.272 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:02:34.272 CC module/bdev/aio/bdev_aio.o 00:02:34.272 CC module/bdev/gpt/gpt.o 00:02:34.272 CC module/bdev/aio/bdev_aio_rpc.o 00:02:34.272 CC module/bdev/gpt/vbdev_gpt.o 00:02:34.272 CC module/bdev/lvol/vbdev_lvol.o 00:02:34.272 CC module/bdev/iscsi/bdev_iscsi.o 00:02:34.272 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:02:34.272 CC module/bdev/null/bdev_null.o 00:02:34.272 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:02:34.272 CC module/bdev/null/bdev_null_rpc.o 00:02:34.272 CC module/bdev/raid/bdev_raid.o 00:02:34.272 CC module/bdev/raid/bdev_raid_rpc.o 00:02:34.272 CC module/bdev/virtio/bdev_virtio_scsi.o 00:02:34.272 CC module/bdev/virtio/bdev_virtio_blk.o 00:02:34.272 CC module/bdev/raid/bdev_raid_sb.o 00:02:34.272 CC module/bdev/virtio/bdev_virtio_rpc.o 00:02:34.272 CC module/bdev/raid/raid0.o 00:02:34.272 CC module/bdev/raid/raid1.o 00:02:34.272 CC module/bdev/raid/concat.o 00:02:34.533 LIB libspdk_blobfs_bdev.a 00:02:34.533 SO libspdk_blobfs_bdev.so.6.0 00:02:34.533 LIB libspdk_bdev_null.a 00:02:34.533 SYMLINK libspdk_blobfs_bdev.so 00:02:34.533 LIB libspdk_bdev_error.a 00:02:34.533 LIB libspdk_bdev_split.a 00:02:34.533 LIB libspdk_bdev_gpt.a 00:02:34.533 SO libspdk_bdev_null.so.6.0 00:02:34.794 LIB libspdk_bdev_ftl.a 00:02:34.794 LIB libspdk_bdev_passthru.a 00:02:34.794 SO libspdk_bdev_error.so.6.0 00:02:34.794 SO libspdk_bdev_split.so.6.0 00:02:34.794 SO libspdk_bdev_gpt.so.6.0 00:02:34.794 SO libspdk_bdev_passthru.so.6.0 00:02:34.794 SO libspdk_bdev_ftl.so.6.0 00:02:34.794 LIB libspdk_bdev_aio.a 00:02:34.794 LIB libspdk_bdev_malloc.a 00:02:34.794 SYMLINK libspdk_bdev_null.so 00:02:34.794 LIB libspdk_bdev_zone_block.a 00:02:34.794 SYMLINK libspdk_bdev_error.so 00:02:34.794 SYMLINK libspdk_bdev_split.so 00:02:34.794 SO libspdk_bdev_aio.so.6.0 00:02:34.794 SYMLINK libspdk_bdev_gpt.so 00:02:34.794 LIB libspdk_bdev_delay.a 00:02:34.794 SO libspdk_bdev_malloc.so.6.0 00:02:34.794 LIB libspdk_bdev_iscsi.a 00:02:34.794 SO libspdk_bdev_zone_block.so.6.0 00:02:34.794 SYMLINK libspdk_bdev_ftl.so 00:02:34.794 SYMLINK libspdk_bdev_passthru.so 00:02:34.794 SO libspdk_bdev_delay.so.6.0 00:02:34.794 SO libspdk_bdev_iscsi.so.6.0 00:02:34.794 SYMLINK libspdk_bdev_aio.so 00:02:34.794 SYMLINK libspdk_bdev_malloc.so 00:02:34.794 SYMLINK libspdk_bdev_zone_block.so 00:02:34.794 LIB libspdk_bdev_lvol.a 00:02:34.794 SYMLINK libspdk_bdev_delay.so 00:02:34.794 SYMLINK libspdk_bdev_iscsi.so 00:02:34.794 LIB libspdk_bdev_virtio.a 00:02:35.054 SO libspdk_bdev_lvol.so.6.0 00:02:35.054 SO libspdk_bdev_virtio.so.6.0 00:02:35.054 SYMLINK libspdk_bdev_lvol.so 00:02:35.054 SYMLINK libspdk_bdev_virtio.so 00:02:35.314 LIB libspdk_bdev_raid.a 00:02:35.314 SO libspdk_bdev_raid.so.6.0 00:02:35.314 SYMLINK libspdk_bdev_raid.so 00:02:36.717 LIB libspdk_bdev_nvme.a 00:02:36.717 SO libspdk_bdev_nvme.so.7.1 00:02:36.717 SYMLINK libspdk_bdev_nvme.so 00:02:37.663 CC module/event/subsystems/iobuf/iobuf.o 00:02:37.663 CC module/event/subsystems/fsdev/fsdev.o 00:02:37.663 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:02:37.663 CC module/event/subsystems/vfu_tgt/vfu_tgt.o 00:02:37.663 CC module/event/subsystems/vmd/vmd.o 00:02:37.663 CC module/event/subsystems/vmd/vmd_rpc.o 00:02:37.663 CC module/event/subsystems/keyring/keyring.o 00:02:37.663 CC module/event/subsystems/sock/sock.o 00:02:37.663 CC module/event/subsystems/scheduler/scheduler.o 00:02:37.663 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:02:37.663 LIB libspdk_event_vfu_tgt.a 00:02:37.663 LIB libspdk_event_fsdev.a 00:02:37.663 LIB libspdk_event_sock.a 00:02:37.663 LIB libspdk_event_keyring.a 00:02:37.663 LIB libspdk_event_iobuf.a 00:02:37.663 LIB libspdk_event_vmd.a 00:02:37.663 LIB libspdk_event_vhost_blk.a 00:02:37.663 SO libspdk_event_vfu_tgt.so.3.0 00:02:37.663 LIB libspdk_event_scheduler.a 00:02:37.663 SO libspdk_event_fsdev.so.1.0 00:02:37.663 SO libspdk_event_sock.so.5.0 00:02:37.663 SO libspdk_event_keyring.so.1.0 00:02:37.663 SO libspdk_event_vhost_blk.so.3.0 00:02:37.663 SO libspdk_event_iobuf.so.3.0 00:02:37.663 SO libspdk_event_vmd.so.6.0 00:02:37.663 SO libspdk_event_scheduler.so.4.0 00:02:37.663 SYMLINK libspdk_event_vfu_tgt.so 00:02:37.663 SYMLINK libspdk_event_fsdev.so 00:02:37.663 SYMLINK libspdk_event_sock.so 00:02:37.663 SYMLINK libspdk_event_keyring.so 00:02:37.663 SYMLINK libspdk_event_vhost_blk.so 00:02:37.663 SYMLINK libspdk_event_iobuf.so 00:02:37.663 SYMLINK libspdk_event_scheduler.so 00:02:37.663 SYMLINK libspdk_event_vmd.so 00:02:38.235 CC module/event/subsystems/accel/accel.o 00:02:38.235 LIB libspdk_event_accel.a 00:02:38.235 SO libspdk_event_accel.so.6.0 00:02:38.235 SYMLINK libspdk_event_accel.so 00:02:38.807 CC module/event/subsystems/bdev/bdev.o 00:02:38.807 LIB libspdk_event_bdev.a 00:02:38.807 SO libspdk_event_bdev.so.6.0 00:02:38.807 SYMLINK libspdk_event_bdev.so 00:02:39.380 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:02:39.380 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:02:39.380 CC module/event/subsystems/nbd/nbd.o 00:02:39.380 CC module/event/subsystems/scsi/scsi.o 00:02:39.380 CC module/event/subsystems/ublk/ublk.o 00:02:39.380 LIB libspdk_event_ublk.a 00:02:39.380 LIB libspdk_event_nbd.a 00:02:39.380 LIB libspdk_event_scsi.a 00:02:39.380 SO libspdk_event_ublk.so.3.0 00:02:39.380 SO libspdk_event_nbd.so.6.0 00:02:39.380 SO libspdk_event_scsi.so.6.0 00:02:39.381 LIB libspdk_event_nvmf.a 00:02:39.641 SYMLINK libspdk_event_nbd.so 00:02:39.642 SYMLINK libspdk_event_ublk.so 00:02:39.642 SO libspdk_event_nvmf.so.6.0 00:02:39.642 SYMLINK libspdk_event_scsi.so 00:02:39.642 SYMLINK libspdk_event_nvmf.so 00:02:39.903 CC module/event/subsystems/iscsi/iscsi.o 00:02:39.903 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:02:40.165 LIB libspdk_event_vhost_scsi.a 00:02:40.165 LIB libspdk_event_iscsi.a 00:02:40.165 SO libspdk_event_vhost_scsi.so.3.0 00:02:40.165 SO libspdk_event_iscsi.so.6.0 00:02:40.165 SYMLINK libspdk_event_vhost_scsi.so 00:02:40.165 SYMLINK libspdk_event_iscsi.so 00:02:40.426 SO libspdk.so.6.0 00:02:40.426 SYMLINK libspdk.so 00:02:40.688 CC test/rpc_client/rpc_client_test.o 00:02:40.688 CXX app/trace/trace.o 00:02:40.688 TEST_HEADER include/spdk/accel.h 00:02:40.688 TEST_HEADER include/spdk/assert.h 00:02:40.688 TEST_HEADER include/spdk/accel_module.h 00:02:40.688 TEST_HEADER include/spdk/barrier.h 00:02:40.688 TEST_HEADER include/spdk/base64.h 00:02:40.688 TEST_HEADER include/spdk/bdev.h 00:02:40.688 TEST_HEADER include/spdk/bdev_zone.h 00:02:40.688 TEST_HEADER include/spdk/bdev_module.h 00:02:40.688 TEST_HEADER include/spdk/bit_pool.h 00:02:40.688 TEST_HEADER include/spdk/bit_array.h 00:02:40.688 CC app/spdk_top/spdk_top.o 00:02:40.688 TEST_HEADER include/spdk/blob_bdev.h 00:02:40.688 TEST_HEADER include/spdk/blobfs.h 00:02:40.688 TEST_HEADER include/spdk/blobfs_bdev.h 00:02:40.688 TEST_HEADER include/spdk/blob.h 00:02:40.688 CC app/spdk_nvme_perf/perf.o 00:02:40.688 TEST_HEADER include/spdk/config.h 00:02:40.688 TEST_HEADER include/spdk/conf.h 00:02:40.688 TEST_HEADER include/spdk/cpuset.h 00:02:40.688 TEST_HEADER include/spdk/crc32.h 00:02:40.688 CC app/trace_record/trace_record.o 00:02:40.688 TEST_HEADER include/spdk/crc16.h 00:02:40.688 CC app/spdk_nvme_identify/identify.o 00:02:40.688 TEST_HEADER include/spdk/dif.h 00:02:40.688 TEST_HEADER include/spdk/crc64.h 00:02:40.688 CC app/spdk_nvme_discover/discovery_aer.o 00:02:40.688 CC app/spdk_lspci/spdk_lspci.o 00:02:40.688 TEST_HEADER include/spdk/dma.h 00:02:40.688 TEST_HEADER include/spdk/endian.h 00:02:40.688 TEST_HEADER include/spdk/env_dpdk.h 00:02:40.688 TEST_HEADER include/spdk/env.h 00:02:40.688 TEST_HEADER include/spdk/fd_group.h 00:02:40.688 TEST_HEADER include/spdk/event.h 00:02:40.688 TEST_HEADER include/spdk/fd.h 00:02:40.688 TEST_HEADER include/spdk/file.h 00:02:40.688 TEST_HEADER include/spdk/fsdev.h 00:02:40.688 TEST_HEADER include/spdk/fsdev_module.h 00:02:40.688 TEST_HEADER include/spdk/fuse_dispatcher.h 00:02:40.688 TEST_HEADER include/spdk/ftl.h 00:02:40.688 TEST_HEADER include/spdk/gpt_spec.h 00:02:40.688 TEST_HEADER include/spdk/hexlify.h 00:02:40.688 TEST_HEADER include/spdk/histogram_data.h 00:02:40.688 TEST_HEADER include/spdk/idxd_spec.h 00:02:40.688 TEST_HEADER include/spdk/idxd.h 00:02:40.688 TEST_HEADER include/spdk/init.h 00:02:40.688 TEST_HEADER include/spdk/iscsi_spec.h 00:02:40.688 TEST_HEADER include/spdk/ioat.h 00:02:40.688 TEST_HEADER include/spdk/ioat_spec.h 00:02:40.688 TEST_HEADER include/spdk/json.h 00:02:40.688 TEST_HEADER include/spdk/jsonrpc.h 00:02:40.688 TEST_HEADER include/spdk/keyring.h 00:02:40.688 TEST_HEADER include/spdk/likely.h 00:02:40.949 TEST_HEADER include/spdk/log.h 00:02:40.949 TEST_HEADER include/spdk/keyring_module.h 00:02:40.949 TEST_HEADER include/spdk/md5.h 00:02:40.949 TEST_HEADER include/spdk/lvol.h 00:02:40.949 CC examples/interrupt_tgt/interrupt_tgt.o 00:02:40.949 TEST_HEADER include/spdk/memory.h 00:02:40.949 TEST_HEADER include/spdk/nbd.h 00:02:40.949 TEST_HEADER include/spdk/mmio.h 00:02:40.949 TEST_HEADER include/spdk/net.h 00:02:40.949 TEST_HEADER include/spdk/notify.h 00:02:40.949 TEST_HEADER include/spdk/nvme.h 00:02:40.949 TEST_HEADER include/spdk/nvme_spec.h 00:02:40.949 TEST_HEADER include/spdk/nvme_intel.h 00:02:40.949 TEST_HEADER include/spdk/nvme_ocssd.h 00:02:40.949 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:02:40.949 TEST_HEADER include/spdk/nvme_zns.h 00:02:40.949 TEST_HEADER include/spdk/nvmf_cmd.h 00:02:40.949 TEST_HEADER include/spdk/nvmf.h 00:02:40.949 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:02:40.949 CC app/spdk_dd/spdk_dd.o 00:02:40.949 TEST_HEADER include/spdk/nvmf_spec.h 00:02:40.949 TEST_HEADER include/spdk/nvmf_transport.h 00:02:40.949 TEST_HEADER include/spdk/pci_ids.h 00:02:40.949 TEST_HEADER include/spdk/opal_spec.h 00:02:40.949 TEST_HEADER include/spdk/opal.h 00:02:40.949 TEST_HEADER include/spdk/queue.h 00:02:40.949 TEST_HEADER include/spdk/pipe.h 00:02:40.949 CC app/iscsi_tgt/iscsi_tgt.o 00:02:40.949 TEST_HEADER include/spdk/reduce.h 00:02:40.949 TEST_HEADER include/spdk/rpc.h 00:02:40.949 TEST_HEADER include/spdk/scheduler.h 00:02:40.949 TEST_HEADER include/spdk/scsi.h 00:02:40.949 TEST_HEADER include/spdk/scsi_spec.h 00:02:40.949 TEST_HEADER include/spdk/sock.h 00:02:40.949 TEST_HEADER include/spdk/stdinc.h 00:02:40.949 TEST_HEADER include/spdk/string.h 00:02:40.949 TEST_HEADER include/spdk/thread.h 00:02:40.949 TEST_HEADER include/spdk/trace.h 00:02:40.949 TEST_HEADER include/spdk/trace_parser.h 00:02:40.949 TEST_HEADER include/spdk/tree.h 00:02:40.949 CC app/nvmf_tgt/nvmf_main.o 00:02:40.949 TEST_HEADER include/spdk/ublk.h 00:02:40.949 TEST_HEADER include/spdk/util.h 00:02:40.949 TEST_HEADER include/spdk/version.h 00:02:40.949 TEST_HEADER include/spdk/uuid.h 00:02:40.949 TEST_HEADER include/spdk/vfio_user_pci.h 00:02:40.949 TEST_HEADER include/spdk/vfio_user_spec.h 00:02:40.949 TEST_HEADER include/spdk/vhost.h 00:02:40.949 TEST_HEADER include/spdk/vmd.h 00:02:40.949 CC app/spdk_tgt/spdk_tgt.o 00:02:40.949 TEST_HEADER include/spdk/xor.h 00:02:40.949 TEST_HEADER include/spdk/zipf.h 00:02:40.949 CXX test/cpp_headers/accel.o 00:02:40.949 CXX test/cpp_headers/accel_module.o 00:02:40.949 CXX test/cpp_headers/assert.o 00:02:40.949 CXX test/cpp_headers/barrier.o 00:02:40.949 CXX test/cpp_headers/base64.o 00:02:40.949 CXX test/cpp_headers/bdev.o 00:02:40.949 CXX test/cpp_headers/bdev_module.o 00:02:40.949 CXX test/cpp_headers/bit_array.o 00:02:40.949 CXX test/cpp_headers/bdev_zone.o 00:02:40.949 CXX test/cpp_headers/bit_pool.o 00:02:40.949 CXX test/cpp_headers/blobfs.o 00:02:40.949 CXX test/cpp_headers/blob_bdev.o 00:02:40.949 CXX test/cpp_headers/blobfs_bdev.o 00:02:40.949 CXX test/cpp_headers/blob.o 00:02:40.949 CXX test/cpp_headers/conf.o 00:02:40.949 CXX test/cpp_headers/cpuset.o 00:02:40.949 CXX test/cpp_headers/config.o 00:02:40.949 CXX test/cpp_headers/crc64.o 00:02:40.949 CXX test/cpp_headers/crc32.o 00:02:40.949 CXX test/cpp_headers/crc16.o 00:02:40.949 CXX test/cpp_headers/dif.o 00:02:40.949 CXX test/cpp_headers/endian.o 00:02:40.949 CXX test/cpp_headers/dma.o 00:02:40.949 CXX test/cpp_headers/env.o 00:02:40.950 CXX test/cpp_headers/event.o 00:02:40.950 CXX test/cpp_headers/env_dpdk.o 00:02:40.950 CXX test/cpp_headers/fd.o 00:02:40.950 CXX test/cpp_headers/fd_group.o 00:02:40.950 CXX test/cpp_headers/file.o 00:02:40.950 CXX test/cpp_headers/fsdev.o 00:02:40.950 CXX test/cpp_headers/fsdev_module.o 00:02:40.950 CXX test/cpp_headers/ftl.o 00:02:40.950 CXX test/cpp_headers/gpt_spec.o 00:02:40.950 CXX test/cpp_headers/hexlify.o 00:02:40.950 CXX test/cpp_headers/fuse_dispatcher.o 00:02:40.950 CXX test/cpp_headers/histogram_data.o 00:02:40.950 CXX test/cpp_headers/idxd.o 00:02:40.950 CXX test/cpp_headers/idxd_spec.o 00:02:40.950 CXX test/cpp_headers/init.o 00:02:40.950 CXX test/cpp_headers/ioat.o 00:02:40.950 CXX test/cpp_headers/iscsi_spec.o 00:02:40.950 CXX test/cpp_headers/jsonrpc.o 00:02:40.950 CXX test/cpp_headers/ioat_spec.o 00:02:40.950 CXX test/cpp_headers/json.o 00:02:40.950 CXX test/cpp_headers/likely.o 00:02:40.950 CXX test/cpp_headers/keyring_module.o 00:02:40.950 CXX test/cpp_headers/keyring.o 00:02:40.950 CXX test/cpp_headers/lvol.o 00:02:40.950 CXX test/cpp_headers/log.o 00:02:40.950 CXX test/cpp_headers/md5.o 00:02:40.950 CXX test/cpp_headers/mmio.o 00:02:40.950 CXX test/cpp_headers/memory.o 00:02:40.950 CXX test/cpp_headers/net.o 00:02:40.950 CXX test/cpp_headers/notify.o 00:02:40.950 CXX test/cpp_headers/nbd.o 00:02:40.950 CXX test/cpp_headers/nvme_intel.o 00:02:40.950 CXX test/cpp_headers/nvme.o 00:02:40.950 CXX test/cpp_headers/nvme_ocssd.o 00:02:40.950 CXX test/cpp_headers/nvme_spec.o 00:02:40.950 CXX test/cpp_headers/nvme_ocssd_spec.o 00:02:40.950 CXX test/cpp_headers/nvme_zns.o 00:02:40.950 CXX test/cpp_headers/nvmf_fc_spec.o 00:02:40.950 CXX test/cpp_headers/nvmf_cmd.o 00:02:40.950 CXX test/cpp_headers/nvmf.o 00:02:40.950 CXX test/cpp_headers/nvmf_transport.o 00:02:40.950 CXX test/cpp_headers/opal.o 00:02:40.950 CXX test/cpp_headers/pci_ids.o 00:02:40.950 CXX test/cpp_headers/nvmf_spec.o 00:02:40.950 CXX test/cpp_headers/opal_spec.o 00:02:40.950 CXX test/cpp_headers/reduce.o 00:02:40.950 CXX test/cpp_headers/pipe.o 00:02:40.950 CXX test/cpp_headers/queue.o 00:02:40.950 CXX test/cpp_headers/rpc.o 00:02:40.950 CXX test/cpp_headers/scsi_spec.o 00:02:40.950 CXX test/cpp_headers/scheduler.o 00:02:40.950 CXX test/cpp_headers/scsi.o 00:02:40.950 CXX test/cpp_headers/sock.o 00:02:40.950 CXX test/cpp_headers/stdinc.o 00:02:40.950 CXX test/cpp_headers/string.o 00:02:40.950 CXX test/cpp_headers/thread.o 00:02:40.950 CXX test/cpp_headers/trace_parser.o 00:02:40.950 CXX test/cpp_headers/trace.o 00:02:40.950 CXX test/cpp_headers/tree.o 00:02:40.950 CXX test/cpp_headers/version.o 00:02:40.950 CXX test/cpp_headers/ublk.o 00:02:40.950 CXX test/cpp_headers/uuid.o 00:02:40.950 CXX test/cpp_headers/util.o 00:02:40.950 CC test/thread/poller_perf/poller_perf.o 00:02:40.950 CXX test/cpp_headers/vfio_user_pci.o 00:02:40.950 CXX test/cpp_headers/zipf.o 00:02:40.950 CC examples/util/zipf/zipf.o 00:02:40.950 CXX test/cpp_headers/vfio_user_spec.o 00:02:40.950 CXX test/cpp_headers/vhost.o 00:02:40.950 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:02:40.950 CXX test/cpp_headers/vmd.o 00:02:40.950 CXX test/cpp_headers/xor.o 00:02:40.950 CC test/env/vtophys/vtophys.o 00:02:40.950 CC test/env/pci/pci_ut.o 00:02:40.950 CC test/env/memory/memory_ut.o 00:02:40.950 CC test/app/histogram_perf/histogram_perf.o 00:02:40.950 CC test/dma/test_dma/test_dma.o 00:02:40.950 CC examples/ioat/verify/verify.o 00:02:41.212 CC test/app/jsoncat/jsoncat.o 00:02:41.212 LINK rpc_client_test 00:02:41.212 CC test/app/stub/stub.o 00:02:41.212 CC examples/ioat/perf/perf.o 00:02:41.212 CC app/fio/nvme/fio_plugin.o 00:02:41.212 LINK spdk_lspci 00:02:41.212 CC test/app/bdev_svc/bdev_svc.o 00:02:41.212 CC app/fio/bdev/fio_plugin.o 00:02:41.212 LINK spdk_nvme_discover 00:02:41.212 LINK spdk_trace_record 00:02:41.212 LINK interrupt_tgt 00:02:41.472 LINK nvmf_tgt 00:02:41.472 LINK spdk_tgt 00:02:41.472 CC test/env/mem_callbacks/mem_callbacks.o 00:02:41.472 LINK zipf 00:02:41.472 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:02:41.472 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:02:41.472 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:02:41.472 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:02:41.472 LINK poller_perf 00:02:41.472 LINK vtophys 00:02:41.472 LINK iscsi_tgt 00:02:41.733 LINK ioat_perf 00:02:41.733 LINK histogram_perf 00:02:41.733 LINK env_dpdk_post_init 00:02:41.733 LINK jsoncat 00:02:41.733 LINK spdk_dd 00:02:41.733 LINK stub 00:02:41.733 LINK bdev_svc 00:02:41.733 LINK spdk_trace 00:02:41.733 LINK verify 00:02:41.994 CC test/event/reactor/reactor.o 00:02:41.994 CC test/event/event_perf/event_perf.o 00:02:41.994 CC test/event/reactor_perf/reactor_perf.o 00:02:41.994 CC examples/idxd/perf/perf.o 00:02:41.994 CC examples/sock/hello_world/hello_sock.o 00:02:41.994 CC test/event/app_repeat/app_repeat.o 00:02:41.994 CC examples/vmd/led/led.o 00:02:41.994 CC examples/vmd/lsvmd/lsvmd.o 00:02:41.994 LINK vhost_fuzz 00:02:41.994 LINK pci_ut 00:02:41.994 CC test/event/scheduler/scheduler.o 00:02:41.994 CC examples/thread/thread/thread_ex.o 00:02:41.994 LINK test_dma 00:02:41.994 LINK nvme_fuzz 00:02:41.994 LINK spdk_bdev 00:02:41.994 LINK reactor_perf 00:02:41.994 LINK event_perf 00:02:41.994 LINK reactor 00:02:41.994 LINK lsvmd 00:02:41.994 LINK mem_callbacks 00:02:41.994 LINK led 00:02:42.255 CC app/vhost/vhost.o 00:02:42.255 LINK spdk_nvme 00:02:42.255 LINK app_repeat 00:02:42.256 LINK spdk_nvme_identify 00:02:42.256 LINK scheduler 00:02:42.256 LINK spdk_top 00:02:42.256 LINK spdk_nvme_perf 00:02:42.256 LINK hello_sock 00:02:42.256 LINK idxd_perf 00:02:42.256 LINK thread 00:02:42.256 LINK vhost 00:02:42.517 CC test/nvme/startup/startup.o 00:02:42.517 CC test/nvme/aer/aer.o 00:02:42.517 CC test/nvme/reset/reset.o 00:02:42.517 CC test/nvme/sgl/sgl.o 00:02:42.517 CC test/nvme/reserve/reserve.o 00:02:42.517 CC test/nvme/boot_partition/boot_partition.o 00:02:42.517 CC test/nvme/connect_stress/connect_stress.o 00:02:42.517 CC test/nvme/fused_ordering/fused_ordering.o 00:02:42.517 CC test/nvme/compliance/nvme_compliance.o 00:02:42.517 CC test/nvme/doorbell_aers/doorbell_aers.o 00:02:42.517 CC test/nvme/cuse/cuse.o 00:02:42.517 CC test/nvme/fdp/fdp.o 00:02:42.517 CC test/nvme/overhead/overhead.o 00:02:42.517 CC test/nvme/err_injection/err_injection.o 00:02:42.517 CC test/nvme/simple_copy/simple_copy.o 00:02:42.517 CC test/nvme/e2edp/nvme_dp.o 00:02:42.517 LINK memory_ut 00:02:42.517 CC test/accel/dif/dif.o 00:02:42.517 CC test/blobfs/mkfs/mkfs.o 00:02:42.777 CC examples/nvme/reconnect/reconnect.o 00:02:42.777 CC examples/nvme/arbitration/arbitration.o 00:02:42.777 CC examples/nvme/hotplug/hotplug.o 00:02:42.777 CC examples/nvme/cmb_copy/cmb_copy.o 00:02:42.777 CC examples/nvme/abort/abort.o 00:02:42.777 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:02:42.777 CC examples/nvme/hello_world/hello_world.o 00:02:42.777 CC examples/nvme/nvme_manage/nvme_manage.o 00:02:42.777 CC test/lvol/esnap/esnap.o 00:02:42.777 LINK startup 00:02:42.777 LINK connect_stress 00:02:42.777 LINK boot_partition 00:02:42.777 LINK err_injection 00:02:42.777 LINK doorbell_aers 00:02:42.777 LINK fused_ordering 00:02:42.777 CC examples/accel/perf/accel_perf.o 00:02:42.777 LINK reserve 00:02:42.777 LINK reset 00:02:42.777 LINK sgl 00:02:42.777 LINK simple_copy 00:02:42.777 LINK aer 00:02:42.777 CC examples/fsdev/hello_world/hello_fsdev.o 00:02:42.777 CC examples/blob/cli/blobcli.o 00:02:42.777 LINK mkfs 00:02:42.777 CC examples/blob/hello_world/hello_blob.o 00:02:42.777 LINK nvme_dp 00:02:42.777 LINK overhead 00:02:43.037 LINK fdp 00:02:43.037 LINK nvme_compliance 00:02:43.037 LINK pmr_persistence 00:02:43.037 LINK cmb_copy 00:02:43.037 LINK hello_world 00:02:43.037 LINK hotplug 00:02:43.037 LINK iscsi_fuzz 00:02:43.037 LINK arbitration 00:02:43.037 LINK reconnect 00:02:43.037 LINK abort 00:02:43.037 LINK hello_blob 00:02:43.037 LINK hello_fsdev 00:02:43.298 LINK nvme_manage 00:02:43.298 LINK dif 00:02:43.298 LINK accel_perf 00:02:43.298 LINK blobcli 00:02:43.868 LINK cuse 00:02:43.868 CC test/bdev/bdevio/bdevio.o 00:02:43.868 CC examples/bdev/bdevperf/bdevperf.o 00:02:43.868 CC examples/bdev/hello_world/hello_bdev.o 00:02:44.128 LINK hello_bdev 00:02:44.128 LINK bdevio 00:02:44.699 LINK bdevperf 00:02:45.270 CC examples/nvmf/nvmf/nvmf.o 00:02:45.530 LINK nvmf 00:02:47.440 LINK esnap 00:02:47.440 00:02:47.440 real 0m54.531s 00:02:47.440 user 7m47.411s 00:02:47.440 sys 4m29.662s 00:02:47.440 18:58:59 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:47.440 18:58:59 make -- common/autotest_common.sh@10 -- $ set +x 00:02:47.440 ************************************ 00:02:47.440 END TEST make 00:02:47.440 ************************************ 00:02:47.440 18:59:00 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:02:47.440 18:59:00 -- pm/common@29 -- $ signal_monitor_resources TERM 00:02:47.440 18:59:00 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:02:47.440 18:59:00 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.440 18:59:00 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:02:47.440 18:59:00 -- pm/common@44 -- $ pid=3512468 00:02:47.440 18:59:00 -- pm/common@50 -- $ kill -TERM 3512468 00:02:47.440 18:59:00 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.440 18:59:00 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:02:47.440 18:59:00 -- pm/common@44 -- $ pid=3512469 00:02:47.440 18:59:00 -- pm/common@50 -- $ kill -TERM 3512469 00:02:47.440 18:59:00 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.440 18:59:00 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:02:47.440 18:59:00 -- pm/common@44 -- $ pid=3512471 00:02:47.440 18:59:00 -- pm/common@50 -- $ kill -TERM 3512471 00:02:47.440 18:59:00 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.440 18:59:00 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:02:47.440 18:59:00 -- pm/common@44 -- $ pid=3512497 00:02:47.440 18:59:00 -- pm/common@50 -- $ sudo -E kill -TERM 3512497 00:02:47.701 18:59:00 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:02:47.701 18:59:00 -- spdk/autorun.sh@27 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:02:47.701 18:59:00 -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:02:47.701 18:59:00 -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:02:47.701 18:59:00 -- common/autotest_common.sh@1693 -- # lcov --version 00:02:47.701 18:59:00 -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:02:47.701 18:59:00 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:02:47.701 18:59:00 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:02:47.701 18:59:00 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:02:47.701 18:59:00 -- scripts/common.sh@336 -- # IFS=.-: 00:02:47.701 18:59:00 -- scripts/common.sh@336 -- # read -ra ver1 00:02:47.701 18:59:00 -- scripts/common.sh@337 -- # IFS=.-: 00:02:47.701 18:59:00 -- scripts/common.sh@337 -- # read -ra ver2 00:02:47.701 18:59:00 -- scripts/common.sh@338 -- # local 'op=<' 00:02:47.701 18:59:00 -- scripts/common.sh@340 -- # ver1_l=2 00:02:47.701 18:59:00 -- scripts/common.sh@341 -- # ver2_l=1 00:02:47.701 18:59:00 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:02:47.701 18:59:00 -- scripts/common.sh@344 -- # case "$op" in 00:02:47.701 18:59:00 -- scripts/common.sh@345 -- # : 1 00:02:47.701 18:59:00 -- scripts/common.sh@364 -- # (( v = 0 )) 00:02:47.701 18:59:00 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:47.701 18:59:00 -- scripts/common.sh@365 -- # decimal 1 00:02:47.701 18:59:00 -- scripts/common.sh@353 -- # local d=1 00:02:47.701 18:59:00 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:02:47.701 18:59:00 -- scripts/common.sh@355 -- # echo 1 00:02:47.701 18:59:00 -- scripts/common.sh@365 -- # ver1[v]=1 00:02:47.701 18:59:00 -- scripts/common.sh@366 -- # decimal 2 00:02:47.701 18:59:00 -- scripts/common.sh@353 -- # local d=2 00:02:47.701 18:59:00 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:02:47.701 18:59:00 -- scripts/common.sh@355 -- # echo 2 00:02:47.701 18:59:00 -- scripts/common.sh@366 -- # ver2[v]=2 00:02:47.701 18:59:00 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:02:47.701 18:59:00 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:02:47.701 18:59:00 -- scripts/common.sh@368 -- # return 0 00:02:47.701 18:59:00 -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:02:47.701 18:59:00 -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:02:47.701 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:47.701 --rc genhtml_branch_coverage=1 00:02:47.701 --rc genhtml_function_coverage=1 00:02:47.701 --rc genhtml_legend=1 00:02:47.702 --rc geninfo_all_blocks=1 00:02:47.702 --rc geninfo_unexecuted_blocks=1 00:02:47.702 00:02:47.702 ' 00:02:47.702 18:59:00 -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:02:47.702 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:47.702 --rc genhtml_branch_coverage=1 00:02:47.702 --rc genhtml_function_coverage=1 00:02:47.702 --rc genhtml_legend=1 00:02:47.702 --rc geninfo_all_blocks=1 00:02:47.702 --rc geninfo_unexecuted_blocks=1 00:02:47.702 00:02:47.702 ' 00:02:47.702 18:59:00 -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:02:47.702 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:47.702 --rc genhtml_branch_coverage=1 00:02:47.702 --rc genhtml_function_coverage=1 00:02:47.702 --rc genhtml_legend=1 00:02:47.702 --rc geninfo_all_blocks=1 00:02:47.702 --rc geninfo_unexecuted_blocks=1 00:02:47.702 00:02:47.702 ' 00:02:47.702 18:59:00 -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:02:47.702 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:47.702 --rc genhtml_branch_coverage=1 00:02:47.702 --rc genhtml_function_coverage=1 00:02:47.702 --rc genhtml_legend=1 00:02:47.702 --rc geninfo_all_blocks=1 00:02:47.702 --rc geninfo_unexecuted_blocks=1 00:02:47.702 00:02:47.702 ' 00:02:47.702 18:59:00 -- spdk/autotest.sh@25 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:02:47.702 18:59:00 -- nvmf/common.sh@7 -- # uname -s 00:02:47.702 18:59:00 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:02:47.702 18:59:00 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:02:47.702 18:59:00 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:02:47.702 18:59:00 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:02:47.702 18:59:00 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:02:47.702 18:59:00 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:02:47.702 18:59:00 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:02:47.702 18:59:00 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:02:47.702 18:59:00 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:02:47.702 18:59:00 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:02:47.702 18:59:00 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:02:47.702 18:59:00 -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:02:47.702 18:59:00 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:02:47.702 18:59:00 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:02:47.702 18:59:00 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:02:47.702 18:59:00 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:02:47.702 18:59:00 -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:02:47.702 18:59:00 -- scripts/common.sh@15 -- # shopt -s extglob 00:02:47.702 18:59:00 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:02:47.702 18:59:00 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:47.702 18:59:00 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:47.702 18:59:00 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:47.702 18:59:00 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:47.702 18:59:00 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:47.702 18:59:00 -- paths/export.sh@5 -- # export PATH 00:02:47.702 18:59:00 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:47.702 18:59:00 -- nvmf/common.sh@51 -- # : 0 00:02:47.702 18:59:00 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:02:47.702 18:59:00 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:02:47.702 18:59:00 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:02:47.702 18:59:00 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:02:47.702 18:59:00 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:02:47.702 18:59:00 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:02:47.702 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:02:47.702 18:59:00 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:02:47.702 18:59:00 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:02:47.702 18:59:00 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:02:47.702 18:59:00 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:02:47.702 18:59:00 -- spdk/autotest.sh@32 -- # uname -s 00:02:47.702 18:59:00 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:02:47.702 18:59:00 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:02:47.702 18:59:00 -- spdk/autotest.sh@34 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:02:47.702 18:59:00 -- spdk/autotest.sh@39 -- # echo '|/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/core-collector.sh %P %s %t' 00:02:47.964 18:59:00 -- spdk/autotest.sh@40 -- # echo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:02:47.964 18:59:00 -- spdk/autotest.sh@44 -- # modprobe nbd 00:02:47.964 18:59:00 -- spdk/autotest.sh@46 -- # type -P udevadm 00:02:47.964 18:59:00 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:02:47.964 18:59:00 -- spdk/autotest.sh@48 -- # udevadm_pid=3577771 00:02:47.964 18:59:00 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:02:47.964 18:59:00 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:02:47.964 18:59:00 -- pm/common@17 -- # local monitor 00:02:47.964 18:59:00 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.964 18:59:00 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.964 18:59:00 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.964 18:59:00 -- pm/common@21 -- # date +%s 00:02:47.964 18:59:00 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:47.964 18:59:00 -- pm/common@21 -- # date +%s 00:02:47.964 18:59:00 -- pm/common@25 -- # sleep 1 00:02:47.964 18:59:00 -- pm/common@21 -- # date +%s 00:02:47.964 18:59:00 -- pm/common@21 -- # date +%s 00:02:47.964 18:59:00 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1732643940 00:02:47.964 18:59:00 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1732643940 00:02:47.964 18:59:00 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1732643940 00:02:47.964 18:59:00 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1732643940 00:02:47.964 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1732643940_collect-cpu-load.pm.log 00:02:47.964 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1732643940_collect-vmstat.pm.log 00:02:47.964 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1732643940_collect-cpu-temp.pm.log 00:02:47.964 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1732643940_collect-bmc-pm.bmc.pm.log 00:02:48.907 18:59:01 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:02:48.907 18:59:01 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:02:48.907 18:59:01 -- common/autotest_common.sh@726 -- # xtrace_disable 00:02:48.907 18:59:01 -- common/autotest_common.sh@10 -- # set +x 00:02:48.907 18:59:01 -- spdk/autotest.sh@59 -- # create_test_list 00:02:48.907 18:59:01 -- common/autotest_common.sh@752 -- # xtrace_disable 00:02:48.907 18:59:01 -- common/autotest_common.sh@10 -- # set +x 00:02:48.907 18:59:01 -- spdk/autotest.sh@61 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh 00:02:48.907 18:59:01 -- spdk/autotest.sh@61 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:48.907 18:59:01 -- spdk/autotest.sh@61 -- # src=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:48.907 18:59:01 -- spdk/autotest.sh@62 -- # out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:02:48.907 18:59:01 -- spdk/autotest.sh@63 -- # cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:48.907 18:59:01 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:02:48.907 18:59:01 -- common/autotest_common.sh@1457 -- # uname 00:02:48.907 18:59:01 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:02:48.907 18:59:01 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:02:48.907 18:59:01 -- common/autotest_common.sh@1477 -- # uname 00:02:48.907 18:59:01 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:02:48.907 18:59:01 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:02:48.907 18:59:01 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:02:48.907 lcov: LCOV version 1.15 00:02:48.907 18:59:01 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info 00:03:15.489 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:03:15.489 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno 00:03:19.694 18:59:31 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:03:19.694 18:59:31 -- common/autotest_common.sh@726 -- # xtrace_disable 00:03:19.694 18:59:31 -- common/autotest_common.sh@10 -- # set +x 00:03:19.694 18:59:31 -- spdk/autotest.sh@78 -- # rm -f 00:03:19.694 18:59:31 -- spdk/autotest.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:03:23.898 0000:80:01.6 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:80:01.7 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:80:01.4 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:80:01.5 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:80:01.2 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:80:01.3 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:80:01.0 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:80:01.1 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:65:00.0 (144d a80a): Already using the nvme driver 00:03:23.898 0000:00:01.6 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:00:01.7 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:00:01.4 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:00:01.5 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:00:01.2 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:00:01.3 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:00:01.0 (8086 0b00): Already using the ioatdma driver 00:03:23.898 0000:00:01.1 (8086 0b00): Already using the ioatdma driver 00:03:23.898 18:59:36 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:03:23.898 18:59:36 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:03:23.898 18:59:36 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:03:23.898 18:59:36 -- common/autotest_common.sh@1658 -- # local nvme bdf 00:03:23.898 18:59:36 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:03:23.898 18:59:36 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme0n1 00:03:23.898 18:59:36 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:03:23.898 18:59:36 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:03:23.898 18:59:36 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:03:23.898 18:59:36 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:03:23.898 18:59:36 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:03:23.898 18:59:36 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:03:23.898 18:59:36 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:03:23.898 18:59:36 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:03:23.898 18:59:36 -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:03:23.898 No valid GPT data, bailing 00:03:23.898 18:59:36 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:03:23.898 18:59:36 -- scripts/common.sh@394 -- # pt= 00:03:23.898 18:59:36 -- scripts/common.sh@395 -- # return 1 00:03:23.898 18:59:36 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:03:23.898 1+0 records in 00:03:23.898 1+0 records out 00:03:23.898 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00482676 s, 217 MB/s 00:03:23.898 18:59:36 -- spdk/autotest.sh@105 -- # sync 00:03:23.898 18:59:36 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:03:23.898 18:59:36 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:03:23.898 18:59:36 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:03:33.898 18:59:44 -- spdk/autotest.sh@111 -- # uname -s 00:03:33.898 18:59:44 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:03:33.898 18:59:44 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:03:33.898 18:59:44 -- spdk/autotest.sh@115 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:03:36.444 Hugepages 00:03:36.444 node hugesize free / total 00:03:36.444 node0 1048576kB 0 / 0 00:03:36.444 node0 2048kB 0 / 0 00:03:36.444 node1 1048576kB 0 / 0 00:03:36.444 node1 2048kB 0 / 0 00:03:36.444 00:03:36.444 Type BDF Vendor Device NUMA Driver Device Block devices 00:03:36.444 I/OAT 0000:00:01.0 8086 0b00 0 ioatdma - - 00:03:36.444 I/OAT 0000:00:01.1 8086 0b00 0 ioatdma - - 00:03:36.444 I/OAT 0000:00:01.2 8086 0b00 0 ioatdma - - 00:03:36.444 I/OAT 0000:00:01.3 8086 0b00 0 ioatdma - - 00:03:36.444 I/OAT 0000:00:01.4 8086 0b00 0 ioatdma - - 00:03:36.444 I/OAT 0000:00:01.5 8086 0b00 0 ioatdma - - 00:03:36.444 I/OAT 0000:00:01.6 8086 0b00 0 ioatdma - - 00:03:36.444 I/OAT 0000:00:01.7 8086 0b00 0 ioatdma - - 00:03:36.444 NVMe 0000:65:00.0 144d a80a 0 nvme nvme0 nvme0n1 00:03:36.444 I/OAT 0000:80:01.0 8086 0b00 1 ioatdma - - 00:03:36.444 I/OAT 0000:80:01.1 8086 0b00 1 ioatdma - - 00:03:36.444 I/OAT 0000:80:01.2 8086 0b00 1 ioatdma - - 00:03:36.444 I/OAT 0000:80:01.3 8086 0b00 1 ioatdma - - 00:03:36.444 I/OAT 0000:80:01.4 8086 0b00 1 ioatdma - - 00:03:36.444 I/OAT 0000:80:01.5 8086 0b00 1 ioatdma - - 00:03:36.444 I/OAT 0000:80:01.6 8086 0b00 1 ioatdma - - 00:03:36.444 I/OAT 0000:80:01.7 8086 0b00 1 ioatdma - - 00:03:36.444 18:59:48 -- spdk/autotest.sh@117 -- # uname -s 00:03:36.444 18:59:48 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:03:36.444 18:59:48 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:03:36.444 18:59:48 -- common/autotest_common.sh@1516 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:40.755 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:03:40.755 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:03:42.150 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:03:42.411 18:59:54 -- common/autotest_common.sh@1517 -- # sleep 1 00:03:43.797 18:59:55 -- common/autotest_common.sh@1518 -- # bdfs=() 00:03:43.797 18:59:55 -- common/autotest_common.sh@1518 -- # local bdfs 00:03:43.797 18:59:55 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:03:43.797 18:59:55 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:03:43.797 18:59:55 -- common/autotest_common.sh@1498 -- # bdfs=() 00:03:43.797 18:59:55 -- common/autotest_common.sh@1498 -- # local bdfs 00:03:43.797 18:59:55 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:03:43.797 18:59:55 -- common/autotest_common.sh@1499 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:03:43.797 18:59:55 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:03:43.797 18:59:56 -- common/autotest_common.sh@1500 -- # (( 1 == 0 )) 00:03:43.797 18:59:56 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:65:00.0 00:03:43.797 18:59:56 -- common/autotest_common.sh@1522 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:03:48.001 Waiting for block devices as requested 00:03:48.001 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:03:48.001 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:03:48.001 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:03:48.001 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:03:48.002 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:03:48.002 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:03:48.002 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:03:48.002 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:03:48.002 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:03:48.263 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:03:48.263 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:03:48.263 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:03:48.263 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:03:48.523 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:03:48.523 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:03:48.523 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:03:48.783 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:03:49.043 19:00:01 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:03:49.043 19:00:01 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:65:00.0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1487 -- # grep 0000:65:00.0/nvme/nvme 00:03:49.043 19:00:01 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 ]] 00:03:49.043 19:00:01 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:03:49.043 19:00:01 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:03:49.043 19:00:01 -- common/autotest_common.sh@1531 -- # grep oacs 00:03:49.043 19:00:01 -- common/autotest_common.sh@1531 -- # oacs=' 0x5f' 00:03:49.043 19:00:01 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:03:49.043 19:00:01 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:03:49.043 19:00:01 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:03:49.043 19:00:01 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:03:49.043 19:00:01 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:03:49.043 19:00:01 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:03:49.043 19:00:01 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:03:49.043 19:00:01 -- common/autotest_common.sh@1543 -- # continue 00:03:49.043 19:00:01 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:03:49.043 19:00:01 -- common/autotest_common.sh@732 -- # xtrace_disable 00:03:49.043 19:00:01 -- common/autotest_common.sh@10 -- # set +x 00:03:49.043 19:00:01 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:03:49.043 19:00:01 -- common/autotest_common.sh@726 -- # xtrace_disable 00:03:49.043 19:00:01 -- common/autotest_common.sh@10 -- # set +x 00:03:49.043 19:00:01 -- spdk/autotest.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:53.247 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:03:53.247 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:03:53.509 19:00:06 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:03:53.509 19:00:06 -- common/autotest_common.sh@732 -- # xtrace_disable 00:03:53.509 19:00:06 -- common/autotest_common.sh@10 -- # set +x 00:03:53.509 19:00:06 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:03:53.509 19:00:06 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:03:53.509 19:00:06 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:03:53.509 19:00:06 -- common/autotest_common.sh@1563 -- # bdfs=() 00:03:53.509 19:00:06 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:03:53.509 19:00:06 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:03:53.509 19:00:06 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:03:53.509 19:00:06 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:03:53.509 19:00:06 -- common/autotest_common.sh@1498 -- # bdfs=() 00:03:53.509 19:00:06 -- common/autotest_common.sh@1498 -- # local bdfs 00:03:53.509 19:00:06 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:03:53.509 19:00:06 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:03:53.509 19:00:06 -- common/autotest_common.sh@1499 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:03:53.770 19:00:06 -- common/autotest_common.sh@1500 -- # (( 1 == 0 )) 00:03:53.770 19:00:06 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:65:00.0 00:03:53.770 19:00:06 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:03:53.770 19:00:06 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:65:00.0/device 00:03:53.770 19:00:06 -- common/autotest_common.sh@1566 -- # device=0xa80a 00:03:53.770 19:00:06 -- common/autotest_common.sh@1567 -- # [[ 0xa80a == \0\x\0\a\5\4 ]] 00:03:53.770 19:00:06 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:03:53.770 19:00:06 -- common/autotest_common.sh@1572 -- # return 0 00:03:53.770 19:00:06 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:03:53.770 19:00:06 -- common/autotest_common.sh@1580 -- # return 0 00:03:53.770 19:00:06 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:03:53.770 19:00:06 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:03:53.770 19:00:06 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:03:53.770 19:00:06 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:03:53.770 19:00:06 -- spdk/autotest.sh@149 -- # timing_enter lib 00:03:53.770 19:00:06 -- common/autotest_common.sh@726 -- # xtrace_disable 00:03:53.770 19:00:06 -- common/autotest_common.sh@10 -- # set +x 00:03:53.770 19:00:06 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:03:53.770 19:00:06 -- spdk/autotest.sh@155 -- # run_test env /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:03:53.770 19:00:06 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:03:53.770 19:00:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:03:53.770 19:00:06 -- common/autotest_common.sh@10 -- # set +x 00:03:53.770 ************************************ 00:03:53.770 START TEST env 00:03:53.770 ************************************ 00:03:53.770 19:00:06 env -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:03:53.770 * Looking for test storage... 00:03:53.770 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env 00:03:53.770 19:00:06 env -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:03:53.770 19:00:06 env -- common/autotest_common.sh@1693 -- # lcov --version 00:03:53.770 19:00:06 env -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:03:54.031 19:00:06 env -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:03:54.031 19:00:06 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:54.031 19:00:06 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:54.031 19:00:06 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:54.032 19:00:06 env -- scripts/common.sh@336 -- # IFS=.-: 00:03:54.032 19:00:06 env -- scripts/common.sh@336 -- # read -ra ver1 00:03:54.032 19:00:06 env -- scripts/common.sh@337 -- # IFS=.-: 00:03:54.032 19:00:06 env -- scripts/common.sh@337 -- # read -ra ver2 00:03:54.032 19:00:06 env -- scripts/common.sh@338 -- # local 'op=<' 00:03:54.032 19:00:06 env -- scripts/common.sh@340 -- # ver1_l=2 00:03:54.032 19:00:06 env -- scripts/common.sh@341 -- # ver2_l=1 00:03:54.032 19:00:06 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:54.032 19:00:06 env -- scripts/common.sh@344 -- # case "$op" in 00:03:54.032 19:00:06 env -- scripts/common.sh@345 -- # : 1 00:03:54.032 19:00:06 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:54.032 19:00:06 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:54.032 19:00:06 env -- scripts/common.sh@365 -- # decimal 1 00:03:54.032 19:00:06 env -- scripts/common.sh@353 -- # local d=1 00:03:54.032 19:00:06 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:54.032 19:00:06 env -- scripts/common.sh@355 -- # echo 1 00:03:54.032 19:00:06 env -- scripts/common.sh@365 -- # ver1[v]=1 00:03:54.032 19:00:06 env -- scripts/common.sh@366 -- # decimal 2 00:03:54.032 19:00:06 env -- scripts/common.sh@353 -- # local d=2 00:03:54.032 19:00:06 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:54.032 19:00:06 env -- scripts/common.sh@355 -- # echo 2 00:03:54.032 19:00:06 env -- scripts/common.sh@366 -- # ver2[v]=2 00:03:54.032 19:00:06 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:54.032 19:00:06 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:54.032 19:00:06 env -- scripts/common.sh@368 -- # return 0 00:03:54.032 19:00:06 env -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:54.032 19:00:06 env -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:03:54.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:54.032 --rc genhtml_branch_coverage=1 00:03:54.032 --rc genhtml_function_coverage=1 00:03:54.032 --rc genhtml_legend=1 00:03:54.032 --rc geninfo_all_blocks=1 00:03:54.032 --rc geninfo_unexecuted_blocks=1 00:03:54.032 00:03:54.032 ' 00:03:54.032 19:00:06 env -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:03:54.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:54.032 --rc genhtml_branch_coverage=1 00:03:54.032 --rc genhtml_function_coverage=1 00:03:54.032 --rc genhtml_legend=1 00:03:54.032 --rc geninfo_all_blocks=1 00:03:54.032 --rc geninfo_unexecuted_blocks=1 00:03:54.032 00:03:54.032 ' 00:03:54.032 19:00:06 env -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:03:54.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:54.032 --rc genhtml_branch_coverage=1 00:03:54.032 --rc genhtml_function_coverage=1 00:03:54.032 --rc genhtml_legend=1 00:03:54.032 --rc geninfo_all_blocks=1 00:03:54.032 --rc geninfo_unexecuted_blocks=1 00:03:54.032 00:03:54.032 ' 00:03:54.032 19:00:06 env -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:03:54.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:54.032 --rc genhtml_branch_coverage=1 00:03:54.032 --rc genhtml_function_coverage=1 00:03:54.032 --rc genhtml_legend=1 00:03:54.032 --rc geninfo_all_blocks=1 00:03:54.032 --rc geninfo_unexecuted_blocks=1 00:03:54.032 00:03:54.032 ' 00:03:54.032 19:00:06 env -- env/env.sh@10 -- # run_test env_memory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:03:54.032 19:00:06 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:03:54.032 19:00:06 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:03:54.032 19:00:06 env -- common/autotest_common.sh@10 -- # set +x 00:03:54.032 ************************************ 00:03:54.032 START TEST env_memory 00:03:54.032 ************************************ 00:03:54.032 19:00:06 env.env_memory -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:03:54.032 00:03:54.032 00:03:54.032 CUnit - A unit testing framework for C - Version 2.1-3 00:03:54.032 http://cunit.sourceforge.net/ 00:03:54.032 00:03:54.032 00:03:54.032 Suite: memory 00:03:54.032 Test: alloc and free memory map ...[2024-11-26 19:00:06.520554] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:03:54.032 passed 00:03:54.032 Test: mem map translation ...[2024-11-26 19:00:06.546048] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:03:54.032 [2024-11-26 19:00:06.546074] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:03:54.032 [2024-11-26 19:00:06.546121] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:03:54.032 [2024-11-26 19:00:06.546128] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:03:54.032 passed 00:03:54.032 Test: mem map registration ...[2024-11-26 19:00:06.601403] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:03:54.032 [2024-11-26 19:00:06.601425] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:03:54.032 passed 00:03:54.294 Test: mem map adjacent registrations ...passed 00:03:54.294 00:03:54.294 Run Summary: Type Total Ran Passed Failed Inactive 00:03:54.294 suites 1 1 n/a 0 0 00:03:54.294 tests 4 4 4 0 0 00:03:54.294 asserts 152 152 152 0 n/a 00:03:54.294 00:03:54.294 Elapsed time = 0.192 seconds 00:03:54.294 00:03:54.294 real 0m0.207s 00:03:54.294 user 0m0.196s 00:03:54.294 sys 0m0.010s 00:03:54.294 19:00:06 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:03:54.294 19:00:06 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:03:54.294 ************************************ 00:03:54.294 END TEST env_memory 00:03:54.294 ************************************ 00:03:54.294 19:00:06 env -- env/env.sh@11 -- # run_test env_vtophys /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:03:54.294 19:00:06 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:03:54.294 19:00:06 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:03:54.294 19:00:06 env -- common/autotest_common.sh@10 -- # set +x 00:03:54.294 ************************************ 00:03:54.294 START TEST env_vtophys 00:03:54.294 ************************************ 00:03:54.294 19:00:06 env.env_vtophys -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:03:54.294 EAL: lib.eal log level changed from notice to debug 00:03:54.294 EAL: Detected lcore 0 as core 0 on socket 0 00:03:54.294 EAL: Detected lcore 1 as core 1 on socket 0 00:03:54.294 EAL: Detected lcore 2 as core 2 on socket 0 00:03:54.294 EAL: Detected lcore 3 as core 3 on socket 0 00:03:54.294 EAL: Detected lcore 4 as core 4 on socket 0 00:03:54.294 EAL: Detected lcore 5 as core 5 on socket 0 00:03:54.294 EAL: Detected lcore 6 as core 6 on socket 0 00:03:54.294 EAL: Detected lcore 7 as core 7 on socket 0 00:03:54.294 EAL: Detected lcore 8 as core 8 on socket 0 00:03:54.294 EAL: Detected lcore 9 as core 9 on socket 0 00:03:54.294 EAL: Detected lcore 10 as core 10 on socket 0 00:03:54.294 EAL: Detected lcore 11 as core 11 on socket 0 00:03:54.294 EAL: Detected lcore 12 as core 12 on socket 0 00:03:54.294 EAL: Detected lcore 13 as core 13 on socket 0 00:03:54.294 EAL: Detected lcore 14 as core 14 on socket 0 00:03:54.294 EAL: Detected lcore 15 as core 15 on socket 0 00:03:54.294 EAL: Detected lcore 16 as core 16 on socket 0 00:03:54.294 EAL: Detected lcore 17 as core 17 on socket 0 00:03:54.294 EAL: Detected lcore 18 as core 18 on socket 0 00:03:54.294 EAL: Detected lcore 19 as core 19 on socket 0 00:03:54.294 EAL: Detected lcore 20 as core 20 on socket 0 00:03:54.294 EAL: Detected lcore 21 as core 21 on socket 0 00:03:54.294 EAL: Detected lcore 22 as core 22 on socket 0 00:03:54.294 EAL: Detected lcore 23 as core 23 on socket 0 00:03:54.294 EAL: Detected lcore 24 as core 24 on socket 0 00:03:54.294 EAL: Detected lcore 25 as core 25 on socket 0 00:03:54.294 EAL: Detected lcore 26 as core 26 on socket 0 00:03:54.294 EAL: Detected lcore 27 as core 27 on socket 0 00:03:54.294 EAL: Detected lcore 28 as core 28 on socket 0 00:03:54.294 EAL: Detected lcore 29 as core 29 on socket 0 00:03:54.294 EAL: Detected lcore 30 as core 30 on socket 0 00:03:54.294 EAL: Detected lcore 31 as core 31 on socket 0 00:03:54.294 EAL: Detected lcore 32 as core 32 on socket 0 00:03:54.294 EAL: Detected lcore 33 as core 33 on socket 0 00:03:54.294 EAL: Detected lcore 34 as core 34 on socket 0 00:03:54.294 EAL: Detected lcore 35 as core 35 on socket 0 00:03:54.294 EAL: Detected lcore 36 as core 0 on socket 1 00:03:54.294 EAL: Detected lcore 37 as core 1 on socket 1 00:03:54.294 EAL: Detected lcore 38 as core 2 on socket 1 00:03:54.294 EAL: Detected lcore 39 as core 3 on socket 1 00:03:54.294 EAL: Detected lcore 40 as core 4 on socket 1 00:03:54.294 EAL: Detected lcore 41 as core 5 on socket 1 00:03:54.294 EAL: Detected lcore 42 as core 6 on socket 1 00:03:54.294 EAL: Detected lcore 43 as core 7 on socket 1 00:03:54.294 EAL: Detected lcore 44 as core 8 on socket 1 00:03:54.294 EAL: Detected lcore 45 as core 9 on socket 1 00:03:54.294 EAL: Detected lcore 46 as core 10 on socket 1 00:03:54.294 EAL: Detected lcore 47 as core 11 on socket 1 00:03:54.294 EAL: Detected lcore 48 as core 12 on socket 1 00:03:54.294 EAL: Detected lcore 49 as core 13 on socket 1 00:03:54.294 EAL: Detected lcore 50 as core 14 on socket 1 00:03:54.294 EAL: Detected lcore 51 as core 15 on socket 1 00:03:54.294 EAL: Detected lcore 52 as core 16 on socket 1 00:03:54.294 EAL: Detected lcore 53 as core 17 on socket 1 00:03:54.294 EAL: Detected lcore 54 as core 18 on socket 1 00:03:54.294 EAL: Detected lcore 55 as core 19 on socket 1 00:03:54.294 EAL: Detected lcore 56 as core 20 on socket 1 00:03:54.294 EAL: Detected lcore 57 as core 21 on socket 1 00:03:54.294 EAL: Detected lcore 58 as core 22 on socket 1 00:03:54.294 EAL: Detected lcore 59 as core 23 on socket 1 00:03:54.294 EAL: Detected lcore 60 as core 24 on socket 1 00:03:54.294 EAL: Detected lcore 61 as core 25 on socket 1 00:03:54.294 EAL: Detected lcore 62 as core 26 on socket 1 00:03:54.294 EAL: Detected lcore 63 as core 27 on socket 1 00:03:54.294 EAL: Detected lcore 64 as core 28 on socket 1 00:03:54.294 EAL: Detected lcore 65 as core 29 on socket 1 00:03:54.294 EAL: Detected lcore 66 as core 30 on socket 1 00:03:54.294 EAL: Detected lcore 67 as core 31 on socket 1 00:03:54.294 EAL: Detected lcore 68 as core 32 on socket 1 00:03:54.294 EAL: Detected lcore 69 as core 33 on socket 1 00:03:54.294 EAL: Detected lcore 70 as core 34 on socket 1 00:03:54.294 EAL: Detected lcore 71 as core 35 on socket 1 00:03:54.294 EAL: Detected lcore 72 as core 0 on socket 0 00:03:54.294 EAL: Detected lcore 73 as core 1 on socket 0 00:03:54.294 EAL: Detected lcore 74 as core 2 on socket 0 00:03:54.294 EAL: Detected lcore 75 as core 3 on socket 0 00:03:54.294 EAL: Detected lcore 76 as core 4 on socket 0 00:03:54.294 EAL: Detected lcore 77 as core 5 on socket 0 00:03:54.294 EAL: Detected lcore 78 as core 6 on socket 0 00:03:54.294 EAL: Detected lcore 79 as core 7 on socket 0 00:03:54.294 EAL: Detected lcore 80 as core 8 on socket 0 00:03:54.294 EAL: Detected lcore 81 as core 9 on socket 0 00:03:54.294 EAL: Detected lcore 82 as core 10 on socket 0 00:03:54.294 EAL: Detected lcore 83 as core 11 on socket 0 00:03:54.294 EAL: Detected lcore 84 as core 12 on socket 0 00:03:54.294 EAL: Detected lcore 85 as core 13 on socket 0 00:03:54.294 EAL: Detected lcore 86 as core 14 on socket 0 00:03:54.294 EAL: Detected lcore 87 as core 15 on socket 0 00:03:54.294 EAL: Detected lcore 88 as core 16 on socket 0 00:03:54.294 EAL: Detected lcore 89 as core 17 on socket 0 00:03:54.294 EAL: Detected lcore 90 as core 18 on socket 0 00:03:54.294 EAL: Detected lcore 91 as core 19 on socket 0 00:03:54.294 EAL: Detected lcore 92 as core 20 on socket 0 00:03:54.294 EAL: Detected lcore 93 as core 21 on socket 0 00:03:54.294 EAL: Detected lcore 94 as core 22 on socket 0 00:03:54.294 EAL: Detected lcore 95 as core 23 on socket 0 00:03:54.294 EAL: Detected lcore 96 as core 24 on socket 0 00:03:54.294 EAL: Detected lcore 97 as core 25 on socket 0 00:03:54.294 EAL: Detected lcore 98 as core 26 on socket 0 00:03:54.294 EAL: Detected lcore 99 as core 27 on socket 0 00:03:54.294 EAL: Detected lcore 100 as core 28 on socket 0 00:03:54.294 EAL: Detected lcore 101 as core 29 on socket 0 00:03:54.294 EAL: Detected lcore 102 as core 30 on socket 0 00:03:54.294 EAL: Detected lcore 103 as core 31 on socket 0 00:03:54.295 EAL: Detected lcore 104 as core 32 on socket 0 00:03:54.295 EAL: Detected lcore 105 as core 33 on socket 0 00:03:54.295 EAL: Detected lcore 106 as core 34 on socket 0 00:03:54.295 EAL: Detected lcore 107 as core 35 on socket 0 00:03:54.295 EAL: Detected lcore 108 as core 0 on socket 1 00:03:54.295 EAL: Detected lcore 109 as core 1 on socket 1 00:03:54.295 EAL: Detected lcore 110 as core 2 on socket 1 00:03:54.295 EAL: Detected lcore 111 as core 3 on socket 1 00:03:54.295 EAL: Detected lcore 112 as core 4 on socket 1 00:03:54.295 EAL: Detected lcore 113 as core 5 on socket 1 00:03:54.295 EAL: Detected lcore 114 as core 6 on socket 1 00:03:54.295 EAL: Detected lcore 115 as core 7 on socket 1 00:03:54.295 EAL: Detected lcore 116 as core 8 on socket 1 00:03:54.295 EAL: Detected lcore 117 as core 9 on socket 1 00:03:54.295 EAL: Detected lcore 118 as core 10 on socket 1 00:03:54.295 EAL: Detected lcore 119 as core 11 on socket 1 00:03:54.295 EAL: Detected lcore 120 as core 12 on socket 1 00:03:54.295 EAL: Detected lcore 121 as core 13 on socket 1 00:03:54.295 EAL: Detected lcore 122 as core 14 on socket 1 00:03:54.295 EAL: Detected lcore 123 as core 15 on socket 1 00:03:54.295 EAL: Detected lcore 124 as core 16 on socket 1 00:03:54.295 EAL: Detected lcore 125 as core 17 on socket 1 00:03:54.295 EAL: Detected lcore 126 as core 18 on socket 1 00:03:54.295 EAL: Detected lcore 127 as core 19 on socket 1 00:03:54.295 EAL: Skipped lcore 128 as core 20 on socket 1 00:03:54.295 EAL: Skipped lcore 129 as core 21 on socket 1 00:03:54.295 EAL: Skipped lcore 130 as core 22 on socket 1 00:03:54.295 EAL: Skipped lcore 131 as core 23 on socket 1 00:03:54.295 EAL: Skipped lcore 132 as core 24 on socket 1 00:03:54.295 EAL: Skipped lcore 133 as core 25 on socket 1 00:03:54.295 EAL: Skipped lcore 134 as core 26 on socket 1 00:03:54.295 EAL: Skipped lcore 135 as core 27 on socket 1 00:03:54.295 EAL: Skipped lcore 136 as core 28 on socket 1 00:03:54.295 EAL: Skipped lcore 137 as core 29 on socket 1 00:03:54.295 EAL: Skipped lcore 138 as core 30 on socket 1 00:03:54.295 EAL: Skipped lcore 139 as core 31 on socket 1 00:03:54.295 EAL: Skipped lcore 140 as core 32 on socket 1 00:03:54.295 EAL: Skipped lcore 141 as core 33 on socket 1 00:03:54.295 EAL: Skipped lcore 142 as core 34 on socket 1 00:03:54.295 EAL: Skipped lcore 143 as core 35 on socket 1 00:03:54.295 EAL: Maximum logical cores by configuration: 128 00:03:54.295 EAL: Detected CPU lcores: 128 00:03:54.295 EAL: Detected NUMA nodes: 2 00:03:54.295 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:03:54.295 EAL: Detected shared linkage of DPDK 00:03:54.295 EAL: No shared files mode enabled, IPC will be disabled 00:03:54.295 EAL: Bus pci wants IOVA as 'DC' 00:03:54.295 EAL: Buses did not request a specific IOVA mode. 00:03:54.295 EAL: IOMMU is available, selecting IOVA as VA mode. 00:03:54.295 EAL: Selected IOVA mode 'VA' 00:03:54.295 EAL: Probing VFIO support... 00:03:54.295 EAL: IOMMU type 1 (Type 1) is supported 00:03:54.295 EAL: IOMMU type 7 (sPAPR) is not supported 00:03:54.295 EAL: IOMMU type 8 (No-IOMMU) is not supported 00:03:54.295 EAL: VFIO support initialized 00:03:54.295 EAL: Ask a virtual area of 0x2e000 bytes 00:03:54.295 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:03:54.295 EAL: Setting up physically contiguous memory... 00:03:54.295 EAL: Setting maximum number of open files to 524288 00:03:54.295 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:03:54.295 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 00:03:54.295 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:03:54.295 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x201000800000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x201000a00000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x201000a00000, size 400000000 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x201400a00000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x201400c00000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x201400c00000, size 400000000 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x201800c00000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x201800e00000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x201800e00000, size 400000000 00:03:54.295 EAL: Ask a virtual area of 0x61000 bytes 00:03:54.295 EAL: Virtual area found at 0x201c00e00000 (size = 0x61000) 00:03:54.295 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:54.295 EAL: Ask a virtual area of 0x400000000 bytes 00:03:54.295 EAL: Virtual area found at 0x201c01000000 (size = 0x400000000) 00:03:54.295 EAL: VA reserved for memseg list at 0x201c01000000, size 400000000 00:03:54.295 EAL: Hugepages will be freed exactly as allocated. 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: TSC frequency is ~2400000 KHz 00:03:54.295 EAL: Main lcore 0 is ready (tid=7ff000575a00;cpuset=[0]) 00:03:54.295 EAL: Trying to obtain current memory policy. 00:03:54.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.295 EAL: Restoring previous memory policy: 0 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was expanded by 2MB 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: No PCI address specified using 'addr=' in: bus=pci 00:03:54.295 EAL: Mem event callback 'spdk:(nil)' registered 00:03:54.295 00:03:54.295 00:03:54.295 CUnit - A unit testing framework for C - Version 2.1-3 00:03:54.295 http://cunit.sourceforge.net/ 00:03:54.295 00:03:54.295 00:03:54.295 Suite: components_suite 00:03:54.295 Test: vtophys_malloc_test ...passed 00:03:54.295 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:03:54.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.295 EAL: Restoring previous memory policy: 4 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was expanded by 4MB 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was shrunk by 4MB 00:03:54.295 EAL: Trying to obtain current memory policy. 00:03:54.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.295 EAL: Restoring previous memory policy: 4 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was expanded by 6MB 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was shrunk by 6MB 00:03:54.295 EAL: Trying to obtain current memory policy. 00:03:54.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.295 EAL: Restoring previous memory policy: 4 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was expanded by 10MB 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was shrunk by 10MB 00:03:54.295 EAL: Trying to obtain current memory policy. 00:03:54.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.295 EAL: Restoring previous memory policy: 4 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was expanded by 18MB 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was shrunk by 18MB 00:03:54.295 EAL: Trying to obtain current memory policy. 00:03:54.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.295 EAL: Restoring previous memory policy: 4 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.295 EAL: request: mp_malloc_sync 00:03:54.295 EAL: No shared files mode enabled, IPC is disabled 00:03:54.295 EAL: Heap on socket 0 was expanded by 34MB 00:03:54.295 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.296 EAL: request: mp_malloc_sync 00:03:54.296 EAL: No shared files mode enabled, IPC is disabled 00:03:54.296 EAL: Heap on socket 0 was shrunk by 34MB 00:03:54.296 EAL: Trying to obtain current memory policy. 00:03:54.296 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.296 EAL: Restoring previous memory policy: 4 00:03:54.296 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.296 EAL: request: mp_malloc_sync 00:03:54.296 EAL: No shared files mode enabled, IPC is disabled 00:03:54.296 EAL: Heap on socket 0 was expanded by 66MB 00:03:54.296 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.556 EAL: request: mp_malloc_sync 00:03:54.556 EAL: No shared files mode enabled, IPC is disabled 00:03:54.556 EAL: Heap on socket 0 was shrunk by 66MB 00:03:54.556 EAL: Trying to obtain current memory policy. 00:03:54.556 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.556 EAL: Restoring previous memory policy: 4 00:03:54.556 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.556 EAL: request: mp_malloc_sync 00:03:54.556 EAL: No shared files mode enabled, IPC is disabled 00:03:54.556 EAL: Heap on socket 0 was expanded by 130MB 00:03:54.556 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.556 EAL: request: mp_malloc_sync 00:03:54.556 EAL: No shared files mode enabled, IPC is disabled 00:03:54.556 EAL: Heap on socket 0 was shrunk by 130MB 00:03:54.556 EAL: Trying to obtain current memory policy. 00:03:54.556 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.556 EAL: Restoring previous memory policy: 4 00:03:54.556 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.556 EAL: request: mp_malloc_sync 00:03:54.556 EAL: No shared files mode enabled, IPC is disabled 00:03:54.556 EAL: Heap on socket 0 was expanded by 258MB 00:03:54.556 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.556 EAL: request: mp_malloc_sync 00:03:54.556 EAL: No shared files mode enabled, IPC is disabled 00:03:54.556 EAL: Heap on socket 0 was shrunk by 258MB 00:03:54.556 EAL: Trying to obtain current memory policy. 00:03:54.556 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.556 EAL: Restoring previous memory policy: 4 00:03:54.557 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.557 EAL: request: mp_malloc_sync 00:03:54.557 EAL: No shared files mode enabled, IPC is disabled 00:03:54.557 EAL: Heap on socket 0 was expanded by 514MB 00:03:54.557 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.817 EAL: request: mp_malloc_sync 00:03:54.817 EAL: No shared files mode enabled, IPC is disabled 00:03:54.817 EAL: Heap on socket 0 was shrunk by 514MB 00:03:54.817 EAL: Trying to obtain current memory policy. 00:03:54.817 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:54.817 EAL: Restoring previous memory policy: 4 00:03:54.817 EAL: Calling mem event callback 'spdk:(nil)' 00:03:54.817 EAL: request: mp_malloc_sync 00:03:54.817 EAL: No shared files mode enabled, IPC is disabled 00:03:54.817 EAL: Heap on socket 0 was expanded by 1026MB 00:03:55.077 EAL: Calling mem event callback 'spdk:(nil)' 00:03:55.077 EAL: request: mp_malloc_sync 00:03:55.077 EAL: No shared files mode enabled, IPC is disabled 00:03:55.077 EAL: Heap on socket 0 was shrunk by 1026MB 00:03:55.077 passed 00:03:55.077 00:03:55.077 Run Summary: Type Total Ran Passed Failed Inactive 00:03:55.077 suites 1 1 n/a 0 0 00:03:55.077 tests 2 2 2 0 0 00:03:55.077 asserts 497 497 497 0 n/a 00:03:55.077 00:03:55.077 Elapsed time = 0.657 seconds 00:03:55.077 EAL: Calling mem event callback 'spdk:(nil)' 00:03:55.077 EAL: request: mp_malloc_sync 00:03:55.077 EAL: No shared files mode enabled, IPC is disabled 00:03:55.077 EAL: Heap on socket 0 was shrunk by 2MB 00:03:55.077 EAL: No shared files mode enabled, IPC is disabled 00:03:55.077 EAL: No shared files mode enabled, IPC is disabled 00:03:55.077 EAL: No shared files mode enabled, IPC is disabled 00:03:55.077 00:03:55.077 real 0m0.813s 00:03:55.077 user 0m0.428s 00:03:55.077 sys 0m0.353s 00:03:55.077 19:00:07 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:03:55.077 19:00:07 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:03:55.077 ************************************ 00:03:55.077 END TEST env_vtophys 00:03:55.077 ************************************ 00:03:55.077 19:00:07 env -- env/env.sh@12 -- # run_test env_pci /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:03:55.077 19:00:07 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:03:55.077 19:00:07 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:03:55.077 19:00:07 env -- common/autotest_common.sh@10 -- # set +x 00:03:55.077 ************************************ 00:03:55.077 START TEST env_pci 00:03:55.077 ************************************ 00:03:55.077 19:00:07 env.env_pci -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:03:55.077 00:03:55.077 00:03:55.077 CUnit - A unit testing framework for C - Version 2.1-3 00:03:55.077 http://cunit.sourceforge.net/ 00:03:55.077 00:03:55.077 00:03:55.077 Suite: pci 00:03:55.077 Test: pci_hook ...[2024-11-26 19:00:07.665158] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 3598291 has claimed it 00:03:55.337 EAL: Cannot find device (10000:00:01.0) 00:03:55.337 EAL: Failed to attach device on primary process 00:03:55.337 passed 00:03:55.337 00:03:55.337 Run Summary: Type Total Ran Passed Failed Inactive 00:03:55.337 suites 1 1 n/a 0 0 00:03:55.337 tests 1 1 1 0 0 00:03:55.337 asserts 25 25 25 0 n/a 00:03:55.337 00:03:55.337 Elapsed time = 0.035 seconds 00:03:55.337 00:03:55.337 real 0m0.057s 00:03:55.337 user 0m0.015s 00:03:55.337 sys 0m0.042s 00:03:55.337 19:00:07 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:03:55.337 19:00:07 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:03:55.337 ************************************ 00:03:55.337 END TEST env_pci 00:03:55.337 ************************************ 00:03:55.337 19:00:07 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:03:55.338 19:00:07 env -- env/env.sh@15 -- # uname 00:03:55.338 19:00:07 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:03:55.338 19:00:07 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:03:55.338 19:00:07 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:03:55.338 19:00:07 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:03:55.338 19:00:07 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:03:55.338 19:00:07 env -- common/autotest_common.sh@10 -- # set +x 00:03:55.338 ************************************ 00:03:55.338 START TEST env_dpdk_post_init 00:03:55.338 ************************************ 00:03:55.338 19:00:07 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:03:55.338 EAL: Detected CPU lcores: 128 00:03:55.338 EAL: Detected NUMA nodes: 2 00:03:55.338 EAL: Detected shared linkage of DPDK 00:03:55.338 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:03:55.338 EAL: Selected IOVA mode 'VA' 00:03:55.338 EAL: VFIO support initialized 00:03:55.338 TELEMETRY: No legacy callbacks, legacy socket not created 00:03:55.338 EAL: Using IOMMU type 1 (Type 1) 00:03:55.598 EAL: Ignore mapping IO port bar(1) 00:03:55.598 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.0 (socket 0) 00:03:55.859 EAL: Ignore mapping IO port bar(1) 00:03:55.859 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.1 (socket 0) 00:03:56.120 EAL: Ignore mapping IO port bar(1) 00:03:56.120 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.2 (socket 0) 00:03:56.120 EAL: Ignore mapping IO port bar(1) 00:03:56.381 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.3 (socket 0) 00:03:56.381 EAL: Ignore mapping IO port bar(1) 00:03:56.641 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.4 (socket 0) 00:03:56.641 EAL: Ignore mapping IO port bar(1) 00:03:56.900 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.5 (socket 0) 00:03:56.900 EAL: Ignore mapping IO port bar(1) 00:03:56.900 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.6 (socket 0) 00:03:57.160 EAL: Ignore mapping IO port bar(1) 00:03:57.160 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.7 (socket 0) 00:03:57.420 EAL: Probe PCI driver: spdk_nvme (144d:a80a) device: 0000:65:00.0 (socket 0) 00:03:57.679 EAL: Ignore mapping IO port bar(1) 00:03:57.679 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.0 (socket 1) 00:03:57.679 EAL: Ignore mapping IO port bar(1) 00:03:57.940 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.1 (socket 1) 00:03:57.940 EAL: Ignore mapping IO port bar(1) 00:03:58.200 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.2 (socket 1) 00:03:58.200 EAL: Ignore mapping IO port bar(1) 00:03:58.460 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.3 (socket 1) 00:03:58.460 EAL: Ignore mapping IO port bar(1) 00:03:58.460 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.4 (socket 1) 00:03:58.720 EAL: Ignore mapping IO port bar(1) 00:03:58.720 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.5 (socket 1) 00:03:58.980 EAL: Ignore mapping IO port bar(1) 00:03:58.980 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.6 (socket 1) 00:03:59.240 EAL: Ignore mapping IO port bar(1) 00:03:59.240 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.7 (socket 1) 00:03:59.240 EAL: Releasing PCI mapped resource for 0000:65:00.0 00:03:59.240 EAL: Calling pci_unmap_resource for 0000:65:00.0 at 0x202001020000 00:03:59.240 Starting DPDK initialization... 00:03:59.240 Starting SPDK post initialization... 00:03:59.240 SPDK NVMe probe 00:03:59.240 Attaching to 0000:65:00.0 00:03:59.240 Attached to 0000:65:00.0 00:03:59.240 Cleaning up... 00:04:01.152 00:04:01.152 real 0m5.742s 00:04:01.152 user 0m0.113s 00:04:01.152 sys 0m0.176s 00:04:01.152 19:00:13 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:01.152 19:00:13 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:04:01.152 ************************************ 00:04:01.152 END TEST env_dpdk_post_init 00:04:01.152 ************************************ 00:04:01.152 19:00:13 env -- env/env.sh@26 -- # uname 00:04:01.152 19:00:13 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:04:01.152 19:00:13 env -- env/env.sh@29 -- # run_test env_mem_callbacks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:04:01.152 19:00:13 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:01.152 19:00:13 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:01.152 19:00:13 env -- common/autotest_common.sh@10 -- # set +x 00:04:01.152 ************************************ 00:04:01.152 START TEST env_mem_callbacks 00:04:01.152 ************************************ 00:04:01.152 19:00:13 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:04:01.152 EAL: Detected CPU lcores: 128 00:04:01.152 EAL: Detected NUMA nodes: 2 00:04:01.152 EAL: Detected shared linkage of DPDK 00:04:01.152 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:04:01.152 EAL: Selected IOVA mode 'VA' 00:04:01.152 EAL: VFIO support initialized 00:04:01.152 TELEMETRY: No legacy callbacks, legacy socket not created 00:04:01.152 00:04:01.152 00:04:01.152 CUnit - A unit testing framework for C - Version 2.1-3 00:04:01.152 http://cunit.sourceforge.net/ 00:04:01.152 00:04:01.152 00:04:01.152 Suite: memory 00:04:01.152 Test: test ... 00:04:01.152 register 0x200000200000 2097152 00:04:01.152 malloc 3145728 00:04:01.152 register 0x200000400000 4194304 00:04:01.152 buf 0x200000500000 len 3145728 PASSED 00:04:01.152 malloc 64 00:04:01.152 buf 0x2000004fff40 len 64 PASSED 00:04:01.152 malloc 4194304 00:04:01.152 register 0x200000800000 6291456 00:04:01.152 buf 0x200000a00000 len 4194304 PASSED 00:04:01.152 free 0x200000500000 3145728 00:04:01.152 free 0x2000004fff40 64 00:04:01.152 unregister 0x200000400000 4194304 PASSED 00:04:01.152 free 0x200000a00000 4194304 00:04:01.152 unregister 0x200000800000 6291456 PASSED 00:04:01.152 malloc 8388608 00:04:01.152 register 0x200000400000 10485760 00:04:01.152 buf 0x200000600000 len 8388608 PASSED 00:04:01.152 free 0x200000600000 8388608 00:04:01.152 unregister 0x200000400000 10485760 PASSED 00:04:01.152 passed 00:04:01.152 00:04:01.152 Run Summary: Type Total Ran Passed Failed Inactive 00:04:01.152 suites 1 1 n/a 0 0 00:04:01.152 tests 1 1 1 0 0 00:04:01.152 asserts 15 15 15 0 n/a 00:04:01.152 00:04:01.152 Elapsed time = 0.008 seconds 00:04:01.152 00:04:01.152 real 0m0.070s 00:04:01.152 user 0m0.024s 00:04:01.152 sys 0m0.045s 00:04:01.152 19:00:13 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:01.152 19:00:13 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:04:01.152 ************************************ 00:04:01.152 END TEST env_mem_callbacks 00:04:01.152 ************************************ 00:04:01.152 00:04:01.152 real 0m7.500s 00:04:01.152 user 0m1.055s 00:04:01.152 sys 0m0.991s 00:04:01.152 19:00:13 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:01.152 19:00:13 env -- common/autotest_common.sh@10 -- # set +x 00:04:01.152 ************************************ 00:04:01.152 END TEST env 00:04:01.152 ************************************ 00:04:01.152 19:00:13 -- spdk/autotest.sh@156 -- # run_test rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:04:01.152 19:00:13 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:01.152 19:00:13 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:01.152 19:00:13 -- common/autotest_common.sh@10 -- # set +x 00:04:01.413 ************************************ 00:04:01.413 START TEST rpc 00:04:01.413 ************************************ 00:04:01.413 19:00:13 rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:04:01.413 * Looking for test storage... 00:04:01.413 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:01.413 19:00:13 rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:01.413 19:00:13 rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:04:01.413 19:00:13 rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:01.413 19:00:13 rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:01.413 19:00:13 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:01.413 19:00:13 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:01.413 19:00:13 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:01.413 19:00:13 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:04:01.413 19:00:13 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:04:01.413 19:00:13 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:04:01.413 19:00:13 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:04:01.413 19:00:13 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:04:01.413 19:00:13 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:04:01.413 19:00:13 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:04:01.413 19:00:13 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:01.413 19:00:13 rpc -- scripts/common.sh@344 -- # case "$op" in 00:04:01.413 19:00:13 rpc -- scripts/common.sh@345 -- # : 1 00:04:01.413 19:00:13 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:01.413 19:00:13 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:01.413 19:00:13 rpc -- scripts/common.sh@365 -- # decimal 1 00:04:01.413 19:00:14 rpc -- scripts/common.sh@353 -- # local d=1 00:04:01.413 19:00:14 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:01.413 19:00:14 rpc -- scripts/common.sh@355 -- # echo 1 00:04:01.413 19:00:14 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:04:01.413 19:00:14 rpc -- scripts/common.sh@366 -- # decimal 2 00:04:01.413 19:00:14 rpc -- scripts/common.sh@353 -- # local d=2 00:04:01.413 19:00:14 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:01.413 19:00:14 rpc -- scripts/common.sh@355 -- # echo 2 00:04:01.413 19:00:14 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:04:01.413 19:00:14 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:01.413 19:00:14 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:01.413 19:00:14 rpc -- scripts/common.sh@368 -- # return 0 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:01.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:01.413 --rc genhtml_branch_coverage=1 00:04:01.413 --rc genhtml_function_coverage=1 00:04:01.413 --rc genhtml_legend=1 00:04:01.413 --rc geninfo_all_blocks=1 00:04:01.413 --rc geninfo_unexecuted_blocks=1 00:04:01.413 00:04:01.413 ' 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:01.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:01.413 --rc genhtml_branch_coverage=1 00:04:01.413 --rc genhtml_function_coverage=1 00:04:01.413 --rc genhtml_legend=1 00:04:01.413 --rc geninfo_all_blocks=1 00:04:01.413 --rc geninfo_unexecuted_blocks=1 00:04:01.413 00:04:01.413 ' 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:01.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:01.413 --rc genhtml_branch_coverage=1 00:04:01.413 --rc genhtml_function_coverage=1 00:04:01.413 --rc genhtml_legend=1 00:04:01.413 --rc geninfo_all_blocks=1 00:04:01.413 --rc geninfo_unexecuted_blocks=1 00:04:01.413 00:04:01.413 ' 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:01.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:01.413 --rc genhtml_branch_coverage=1 00:04:01.413 --rc genhtml_function_coverage=1 00:04:01.413 --rc genhtml_legend=1 00:04:01.413 --rc geninfo_all_blocks=1 00:04:01.413 --rc geninfo_unexecuted_blocks=1 00:04:01.413 00:04:01.413 ' 00:04:01.413 19:00:14 rpc -- rpc/rpc.sh@65 -- # spdk_pid=3600186 00:04:01.413 19:00:14 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:01.413 19:00:14 rpc -- rpc/rpc.sh@67 -- # waitforlisten 3600186 00:04:01.413 19:00:14 rpc -- rpc/rpc.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -e bdev 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@835 -- # '[' -z 3600186 ']' 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:01.413 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:01.413 19:00:14 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:01.673 [2024-11-26 19:00:14.075422] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:01.673 [2024-11-26 19:00:14.075476] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3600186 ] 00:04:01.673 [2024-11-26 19:00:14.154194] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:01.673 [2024-11-26 19:00:14.189304] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:04:01.673 [2024-11-26 19:00:14.189338] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 3600186' to capture a snapshot of events at runtime. 00:04:01.673 [2024-11-26 19:00:14.189346] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:04:01.673 [2024-11-26 19:00:14.189353] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:04:01.673 [2024-11-26 19:00:14.189359] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid3600186 for offline analysis/debug. 00:04:01.673 [2024-11-26 19:00:14.189915] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:02.614 19:00:14 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:02.614 19:00:14 rpc -- common/autotest_common.sh@868 -- # return 0 00:04:02.614 19:00:14 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:02.614 19:00:14 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:02.614 19:00:14 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:04:02.614 19:00:14 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:04:02.614 19:00:14 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:02.614 19:00:14 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:02.614 19:00:14 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:02.614 ************************************ 00:04:02.614 START TEST rpc_integrity 00:04:02.614 ************************************ 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.614 19:00:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.614 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:04:02.614 { 00:04:02.614 "name": "Malloc0", 00:04:02.614 "aliases": [ 00:04:02.614 "204e17dc-a85c-474f-857a-bf7416739927" 00:04:02.614 ], 00:04:02.614 "product_name": "Malloc disk", 00:04:02.614 "block_size": 512, 00:04:02.614 "num_blocks": 16384, 00:04:02.614 "uuid": "204e17dc-a85c-474f-857a-bf7416739927", 00:04:02.614 "assigned_rate_limits": { 00:04:02.614 "rw_ios_per_sec": 0, 00:04:02.614 "rw_mbytes_per_sec": 0, 00:04:02.614 "r_mbytes_per_sec": 0, 00:04:02.614 "w_mbytes_per_sec": 0 00:04:02.614 }, 00:04:02.614 "claimed": false, 00:04:02.614 "zoned": false, 00:04:02.614 "supported_io_types": { 00:04:02.614 "read": true, 00:04:02.614 "write": true, 00:04:02.614 "unmap": true, 00:04:02.614 "flush": true, 00:04:02.614 "reset": true, 00:04:02.614 "nvme_admin": false, 00:04:02.614 "nvme_io": false, 00:04:02.614 "nvme_io_md": false, 00:04:02.614 "write_zeroes": true, 00:04:02.614 "zcopy": true, 00:04:02.614 "get_zone_info": false, 00:04:02.614 "zone_management": false, 00:04:02.614 "zone_append": false, 00:04:02.614 "compare": false, 00:04:02.614 "compare_and_write": false, 00:04:02.614 "abort": true, 00:04:02.614 "seek_hole": false, 00:04:02.614 "seek_data": false, 00:04:02.614 "copy": true, 00:04:02.614 "nvme_iov_md": false 00:04:02.614 }, 00:04:02.614 "memory_domains": [ 00:04:02.614 { 00:04:02.614 "dma_device_id": "system", 00:04:02.614 "dma_device_type": 1 00:04:02.614 }, 00:04:02.614 { 00:04:02.614 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:02.614 "dma_device_type": 2 00:04:02.614 } 00:04:02.614 ], 00:04:02.614 "driver_specific": {} 00:04:02.615 } 00:04:02.615 ]' 00:04:02.615 19:00:14 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.615 [2024-11-26 19:00:15.036142] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:04:02.615 [2024-11-26 19:00:15.036175] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:04:02.615 [2024-11-26 19:00:15.036188] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1099580 00:04:02.615 [2024-11-26 19:00:15.036195] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:04:02.615 [2024-11-26 19:00:15.037566] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:04:02.615 [2024-11-26 19:00:15.037587] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:04:02.615 Passthru0 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:04:02.615 { 00:04:02.615 "name": "Malloc0", 00:04:02.615 "aliases": [ 00:04:02.615 "204e17dc-a85c-474f-857a-bf7416739927" 00:04:02.615 ], 00:04:02.615 "product_name": "Malloc disk", 00:04:02.615 "block_size": 512, 00:04:02.615 "num_blocks": 16384, 00:04:02.615 "uuid": "204e17dc-a85c-474f-857a-bf7416739927", 00:04:02.615 "assigned_rate_limits": { 00:04:02.615 "rw_ios_per_sec": 0, 00:04:02.615 "rw_mbytes_per_sec": 0, 00:04:02.615 "r_mbytes_per_sec": 0, 00:04:02.615 "w_mbytes_per_sec": 0 00:04:02.615 }, 00:04:02.615 "claimed": true, 00:04:02.615 "claim_type": "exclusive_write", 00:04:02.615 "zoned": false, 00:04:02.615 "supported_io_types": { 00:04:02.615 "read": true, 00:04:02.615 "write": true, 00:04:02.615 "unmap": true, 00:04:02.615 "flush": true, 00:04:02.615 "reset": true, 00:04:02.615 "nvme_admin": false, 00:04:02.615 "nvme_io": false, 00:04:02.615 "nvme_io_md": false, 00:04:02.615 "write_zeroes": true, 00:04:02.615 "zcopy": true, 00:04:02.615 "get_zone_info": false, 00:04:02.615 "zone_management": false, 00:04:02.615 "zone_append": false, 00:04:02.615 "compare": false, 00:04:02.615 "compare_and_write": false, 00:04:02.615 "abort": true, 00:04:02.615 "seek_hole": false, 00:04:02.615 "seek_data": false, 00:04:02.615 "copy": true, 00:04:02.615 "nvme_iov_md": false 00:04:02.615 }, 00:04:02.615 "memory_domains": [ 00:04:02.615 { 00:04:02.615 "dma_device_id": "system", 00:04:02.615 "dma_device_type": 1 00:04:02.615 }, 00:04:02.615 { 00:04:02.615 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:02.615 "dma_device_type": 2 00:04:02.615 } 00:04:02.615 ], 00:04:02.615 "driver_specific": {} 00:04:02.615 }, 00:04:02.615 { 00:04:02.615 "name": "Passthru0", 00:04:02.615 "aliases": [ 00:04:02.615 "2d9548b8-47be-57ca-9b80-c013fa0a50a5" 00:04:02.615 ], 00:04:02.615 "product_name": "passthru", 00:04:02.615 "block_size": 512, 00:04:02.615 "num_blocks": 16384, 00:04:02.615 "uuid": "2d9548b8-47be-57ca-9b80-c013fa0a50a5", 00:04:02.615 "assigned_rate_limits": { 00:04:02.615 "rw_ios_per_sec": 0, 00:04:02.615 "rw_mbytes_per_sec": 0, 00:04:02.615 "r_mbytes_per_sec": 0, 00:04:02.615 "w_mbytes_per_sec": 0 00:04:02.615 }, 00:04:02.615 "claimed": false, 00:04:02.615 "zoned": false, 00:04:02.615 "supported_io_types": { 00:04:02.615 "read": true, 00:04:02.615 "write": true, 00:04:02.615 "unmap": true, 00:04:02.615 "flush": true, 00:04:02.615 "reset": true, 00:04:02.615 "nvme_admin": false, 00:04:02.615 "nvme_io": false, 00:04:02.615 "nvme_io_md": false, 00:04:02.615 "write_zeroes": true, 00:04:02.615 "zcopy": true, 00:04:02.615 "get_zone_info": false, 00:04:02.615 "zone_management": false, 00:04:02.615 "zone_append": false, 00:04:02.615 "compare": false, 00:04:02.615 "compare_and_write": false, 00:04:02.615 "abort": true, 00:04:02.615 "seek_hole": false, 00:04:02.615 "seek_data": false, 00:04:02.615 "copy": true, 00:04:02.615 "nvme_iov_md": false 00:04:02.615 }, 00:04:02.615 "memory_domains": [ 00:04:02.615 { 00:04:02.615 "dma_device_id": "system", 00:04:02.615 "dma_device_type": 1 00:04:02.615 }, 00:04:02.615 { 00:04:02.615 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:02.615 "dma_device_type": 2 00:04:02.615 } 00:04:02.615 ], 00:04:02.615 "driver_specific": { 00:04:02.615 "passthru": { 00:04:02.615 "name": "Passthru0", 00:04:02.615 "base_bdev_name": "Malloc0" 00:04:02.615 } 00:04:02.615 } 00:04:02.615 } 00:04:02.615 ]' 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:04:02.615 19:00:15 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:04:02.615 00:04:02.615 real 0m0.291s 00:04:02.615 user 0m0.176s 00:04:02.615 sys 0m0.047s 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:02.615 19:00:15 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:02.615 ************************************ 00:04:02.615 END TEST rpc_integrity 00:04:02.615 ************************************ 00:04:02.615 19:00:15 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:04:02.615 19:00:15 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:02.615 19:00:15 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:02.615 19:00:15 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:02.876 ************************************ 00:04:02.876 START TEST rpc_plugins 00:04:02.876 ************************************ 00:04:02.876 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:04:02.876 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:04:02.876 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.876 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:02.876 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.876 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:04:02.876 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:04:02.876 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.876 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:02.876 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.876 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:04:02.876 { 00:04:02.876 "name": "Malloc1", 00:04:02.876 "aliases": [ 00:04:02.876 "75a35e4f-c6fe-4a53-b846-031820eb3edc" 00:04:02.876 ], 00:04:02.876 "product_name": "Malloc disk", 00:04:02.876 "block_size": 4096, 00:04:02.876 "num_blocks": 256, 00:04:02.876 "uuid": "75a35e4f-c6fe-4a53-b846-031820eb3edc", 00:04:02.876 "assigned_rate_limits": { 00:04:02.876 "rw_ios_per_sec": 0, 00:04:02.876 "rw_mbytes_per_sec": 0, 00:04:02.876 "r_mbytes_per_sec": 0, 00:04:02.876 "w_mbytes_per_sec": 0 00:04:02.876 }, 00:04:02.876 "claimed": false, 00:04:02.876 "zoned": false, 00:04:02.876 "supported_io_types": { 00:04:02.876 "read": true, 00:04:02.876 "write": true, 00:04:02.876 "unmap": true, 00:04:02.876 "flush": true, 00:04:02.877 "reset": true, 00:04:02.877 "nvme_admin": false, 00:04:02.877 "nvme_io": false, 00:04:02.877 "nvme_io_md": false, 00:04:02.877 "write_zeroes": true, 00:04:02.877 "zcopy": true, 00:04:02.877 "get_zone_info": false, 00:04:02.877 "zone_management": false, 00:04:02.877 "zone_append": false, 00:04:02.877 "compare": false, 00:04:02.877 "compare_and_write": false, 00:04:02.877 "abort": true, 00:04:02.877 "seek_hole": false, 00:04:02.877 "seek_data": false, 00:04:02.877 "copy": true, 00:04:02.877 "nvme_iov_md": false 00:04:02.877 }, 00:04:02.877 "memory_domains": [ 00:04:02.877 { 00:04:02.877 "dma_device_id": "system", 00:04:02.877 "dma_device_type": 1 00:04:02.877 }, 00:04:02.877 { 00:04:02.877 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:02.877 "dma_device_type": 2 00:04:02.877 } 00:04:02.877 ], 00:04:02.877 "driver_specific": {} 00:04:02.877 } 00:04:02.877 ]' 00:04:02.877 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:04:02.877 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:04:02.877 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.877 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:02.877 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:04:02.877 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:04:02.877 19:00:15 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:04:02.877 00:04:02.877 real 0m0.155s 00:04:02.877 user 0m0.097s 00:04:02.877 sys 0m0.022s 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:02.877 19:00:15 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:02.877 ************************************ 00:04:02.877 END TEST rpc_plugins 00:04:02.877 ************************************ 00:04:02.877 19:00:15 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:04:02.877 19:00:15 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:02.877 19:00:15 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:02.877 19:00:15 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:03.138 ************************************ 00:04:03.138 START TEST rpc_trace_cmd_test 00:04:03.138 ************************************ 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:04:03.138 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid3600186", 00:04:03.138 "tpoint_group_mask": "0x8", 00:04:03.138 "iscsi_conn": { 00:04:03.138 "mask": "0x2", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "scsi": { 00:04:03.138 "mask": "0x4", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "bdev": { 00:04:03.138 "mask": "0x8", 00:04:03.138 "tpoint_mask": "0xffffffffffffffff" 00:04:03.138 }, 00:04:03.138 "nvmf_rdma": { 00:04:03.138 "mask": "0x10", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "nvmf_tcp": { 00:04:03.138 "mask": "0x20", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "ftl": { 00:04:03.138 "mask": "0x40", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "blobfs": { 00:04:03.138 "mask": "0x80", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "dsa": { 00:04:03.138 "mask": "0x200", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "thread": { 00:04:03.138 "mask": "0x400", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "nvme_pcie": { 00:04:03.138 "mask": "0x800", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "iaa": { 00:04:03.138 "mask": "0x1000", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "nvme_tcp": { 00:04:03.138 "mask": "0x2000", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "bdev_nvme": { 00:04:03.138 "mask": "0x4000", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "sock": { 00:04:03.138 "mask": "0x8000", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "blob": { 00:04:03.138 "mask": "0x10000", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "bdev_raid": { 00:04:03.138 "mask": "0x20000", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 }, 00:04:03.138 "scheduler": { 00:04:03.138 "mask": "0x40000", 00:04:03.138 "tpoint_mask": "0x0" 00:04:03.138 } 00:04:03.138 }' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:04:03.138 00:04:03.138 real 0m0.242s 00:04:03.138 user 0m0.206s 00:04:03.138 sys 0m0.028s 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:03.138 19:00:15 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:04:03.138 ************************************ 00:04:03.138 END TEST rpc_trace_cmd_test 00:04:03.138 ************************************ 00:04:03.399 19:00:15 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:04:03.399 19:00:15 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:04:03.399 19:00:15 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:04:03.399 19:00:15 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:03.399 19:00:15 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:03.399 19:00:15 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:03.399 ************************************ 00:04:03.399 START TEST rpc_daemon_integrity 00:04:03.399 ************************************ 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.399 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:04:03.399 { 00:04:03.399 "name": "Malloc2", 00:04:03.399 "aliases": [ 00:04:03.399 "c8f7ee3a-5704-4691-877b-16d61d77c6a9" 00:04:03.399 ], 00:04:03.400 "product_name": "Malloc disk", 00:04:03.400 "block_size": 512, 00:04:03.400 "num_blocks": 16384, 00:04:03.400 "uuid": "c8f7ee3a-5704-4691-877b-16d61d77c6a9", 00:04:03.400 "assigned_rate_limits": { 00:04:03.400 "rw_ios_per_sec": 0, 00:04:03.400 "rw_mbytes_per_sec": 0, 00:04:03.400 "r_mbytes_per_sec": 0, 00:04:03.400 "w_mbytes_per_sec": 0 00:04:03.400 }, 00:04:03.400 "claimed": false, 00:04:03.400 "zoned": false, 00:04:03.400 "supported_io_types": { 00:04:03.400 "read": true, 00:04:03.400 "write": true, 00:04:03.400 "unmap": true, 00:04:03.400 "flush": true, 00:04:03.400 "reset": true, 00:04:03.400 "nvme_admin": false, 00:04:03.400 "nvme_io": false, 00:04:03.400 "nvme_io_md": false, 00:04:03.400 "write_zeroes": true, 00:04:03.400 "zcopy": true, 00:04:03.400 "get_zone_info": false, 00:04:03.400 "zone_management": false, 00:04:03.400 "zone_append": false, 00:04:03.400 "compare": false, 00:04:03.400 "compare_and_write": false, 00:04:03.400 "abort": true, 00:04:03.400 "seek_hole": false, 00:04:03.400 "seek_data": false, 00:04:03.400 "copy": true, 00:04:03.400 "nvme_iov_md": false 00:04:03.400 }, 00:04:03.400 "memory_domains": [ 00:04:03.400 { 00:04:03.400 "dma_device_id": "system", 00:04:03.400 "dma_device_type": 1 00:04:03.400 }, 00:04:03.400 { 00:04:03.400 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:03.400 "dma_device_type": 2 00:04:03.400 } 00:04:03.400 ], 00:04:03.400 "driver_specific": {} 00:04:03.400 } 00:04:03.400 ]' 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.400 [2024-11-26 19:00:15.954721] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:04:03.400 [2024-11-26 19:00:15.954751] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:04:03.400 [2024-11-26 19:00:15.954763] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xfe6c50 00:04:03.400 [2024-11-26 19:00:15.954770] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:04:03.400 [2024-11-26 19:00:15.956057] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:04:03.400 [2024-11-26 19:00:15.956077] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:04:03.400 Passthru0 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.400 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:04:03.400 { 00:04:03.400 "name": "Malloc2", 00:04:03.400 "aliases": [ 00:04:03.400 "c8f7ee3a-5704-4691-877b-16d61d77c6a9" 00:04:03.400 ], 00:04:03.400 "product_name": "Malloc disk", 00:04:03.400 "block_size": 512, 00:04:03.400 "num_blocks": 16384, 00:04:03.400 "uuid": "c8f7ee3a-5704-4691-877b-16d61d77c6a9", 00:04:03.400 "assigned_rate_limits": { 00:04:03.400 "rw_ios_per_sec": 0, 00:04:03.400 "rw_mbytes_per_sec": 0, 00:04:03.400 "r_mbytes_per_sec": 0, 00:04:03.400 "w_mbytes_per_sec": 0 00:04:03.400 }, 00:04:03.400 "claimed": true, 00:04:03.400 "claim_type": "exclusive_write", 00:04:03.400 "zoned": false, 00:04:03.400 "supported_io_types": { 00:04:03.400 "read": true, 00:04:03.400 "write": true, 00:04:03.400 "unmap": true, 00:04:03.400 "flush": true, 00:04:03.400 "reset": true, 00:04:03.400 "nvme_admin": false, 00:04:03.400 "nvme_io": false, 00:04:03.400 "nvme_io_md": false, 00:04:03.400 "write_zeroes": true, 00:04:03.400 "zcopy": true, 00:04:03.400 "get_zone_info": false, 00:04:03.400 "zone_management": false, 00:04:03.400 "zone_append": false, 00:04:03.400 "compare": false, 00:04:03.400 "compare_and_write": false, 00:04:03.400 "abort": true, 00:04:03.400 "seek_hole": false, 00:04:03.400 "seek_data": false, 00:04:03.400 "copy": true, 00:04:03.400 "nvme_iov_md": false 00:04:03.400 }, 00:04:03.400 "memory_domains": [ 00:04:03.400 { 00:04:03.400 "dma_device_id": "system", 00:04:03.400 "dma_device_type": 1 00:04:03.400 }, 00:04:03.400 { 00:04:03.400 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:03.400 "dma_device_type": 2 00:04:03.400 } 00:04:03.400 ], 00:04:03.400 "driver_specific": {} 00:04:03.400 }, 00:04:03.400 { 00:04:03.400 "name": "Passthru0", 00:04:03.400 "aliases": [ 00:04:03.400 "61cab1b5-2903-563a-a0d8-92465c7265b0" 00:04:03.400 ], 00:04:03.400 "product_name": "passthru", 00:04:03.400 "block_size": 512, 00:04:03.400 "num_blocks": 16384, 00:04:03.400 "uuid": "61cab1b5-2903-563a-a0d8-92465c7265b0", 00:04:03.400 "assigned_rate_limits": { 00:04:03.400 "rw_ios_per_sec": 0, 00:04:03.400 "rw_mbytes_per_sec": 0, 00:04:03.400 "r_mbytes_per_sec": 0, 00:04:03.400 "w_mbytes_per_sec": 0 00:04:03.400 }, 00:04:03.400 "claimed": false, 00:04:03.400 "zoned": false, 00:04:03.400 "supported_io_types": { 00:04:03.400 "read": true, 00:04:03.400 "write": true, 00:04:03.400 "unmap": true, 00:04:03.400 "flush": true, 00:04:03.400 "reset": true, 00:04:03.400 "nvme_admin": false, 00:04:03.400 "nvme_io": false, 00:04:03.400 "nvme_io_md": false, 00:04:03.400 "write_zeroes": true, 00:04:03.400 "zcopy": true, 00:04:03.400 "get_zone_info": false, 00:04:03.400 "zone_management": false, 00:04:03.400 "zone_append": false, 00:04:03.400 "compare": false, 00:04:03.401 "compare_and_write": false, 00:04:03.401 "abort": true, 00:04:03.401 "seek_hole": false, 00:04:03.401 "seek_data": false, 00:04:03.401 "copy": true, 00:04:03.401 "nvme_iov_md": false 00:04:03.401 }, 00:04:03.401 "memory_domains": [ 00:04:03.401 { 00:04:03.401 "dma_device_id": "system", 00:04:03.401 "dma_device_type": 1 00:04:03.401 }, 00:04:03.401 { 00:04:03.401 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:03.401 "dma_device_type": 2 00:04:03.401 } 00:04:03.401 ], 00:04:03.401 "driver_specific": { 00:04:03.401 "passthru": { 00:04:03.401 "name": "Passthru0", 00:04:03.401 "base_bdev_name": "Malloc2" 00:04:03.401 } 00:04:03.401 } 00:04:03.401 } 00:04:03.401 ]' 00:04:03.401 19:00:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:04:03.661 00:04:03.661 real 0m0.286s 00:04:03.661 user 0m0.181s 00:04:03.661 sys 0m0.040s 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:03.661 19:00:16 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:03.661 ************************************ 00:04:03.661 END TEST rpc_daemon_integrity 00:04:03.661 ************************************ 00:04:03.661 19:00:16 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:04:03.661 19:00:16 rpc -- rpc/rpc.sh@84 -- # killprocess 3600186 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@954 -- # '[' -z 3600186 ']' 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@958 -- # kill -0 3600186 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@959 -- # uname 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3600186 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3600186' 00:04:03.661 killing process with pid 3600186 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@973 -- # kill 3600186 00:04:03.661 19:00:16 rpc -- common/autotest_common.sh@978 -- # wait 3600186 00:04:03.921 00:04:03.921 real 0m2.607s 00:04:03.921 user 0m3.394s 00:04:03.921 sys 0m0.728s 00:04:03.921 19:00:16 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:03.921 19:00:16 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:03.921 ************************************ 00:04:03.921 END TEST rpc 00:04:03.921 ************************************ 00:04:03.922 19:00:16 -- spdk/autotest.sh@157 -- # run_test skip_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:04:03.922 19:00:16 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:03.922 19:00:16 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:03.922 19:00:16 -- common/autotest_common.sh@10 -- # set +x 00:04:03.922 ************************************ 00:04:03.922 START TEST skip_rpc 00:04:03.922 ************************************ 00:04:03.922 19:00:16 skip_rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:04:04.183 * Looking for test storage... 00:04:04.183 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@345 -- # : 1 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:04.183 19:00:16 skip_rpc -- scripts/common.sh@368 -- # return 0 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:04.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:04.183 --rc genhtml_branch_coverage=1 00:04:04.183 --rc genhtml_function_coverage=1 00:04:04.183 --rc genhtml_legend=1 00:04:04.183 --rc geninfo_all_blocks=1 00:04:04.183 --rc geninfo_unexecuted_blocks=1 00:04:04.183 00:04:04.183 ' 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:04.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:04.183 --rc genhtml_branch_coverage=1 00:04:04.183 --rc genhtml_function_coverage=1 00:04:04.183 --rc genhtml_legend=1 00:04:04.183 --rc geninfo_all_blocks=1 00:04:04.183 --rc geninfo_unexecuted_blocks=1 00:04:04.183 00:04:04.183 ' 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:04.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:04.183 --rc genhtml_branch_coverage=1 00:04:04.183 --rc genhtml_function_coverage=1 00:04:04.183 --rc genhtml_legend=1 00:04:04.183 --rc geninfo_all_blocks=1 00:04:04.183 --rc geninfo_unexecuted_blocks=1 00:04:04.183 00:04:04.183 ' 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:04.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:04.183 --rc genhtml_branch_coverage=1 00:04:04.183 --rc genhtml_function_coverage=1 00:04:04.183 --rc genhtml_legend=1 00:04:04.183 --rc geninfo_all_blocks=1 00:04:04.183 --rc geninfo_unexecuted_blocks=1 00:04:04.183 00:04:04.183 ' 00:04:04.183 19:00:16 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:04.183 19:00:16 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:04:04.183 19:00:16 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:04.183 19:00:16 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:04.183 ************************************ 00:04:04.183 START TEST skip_rpc 00:04:04.183 ************************************ 00:04:04.183 19:00:16 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:04:04.183 19:00:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=3600733 00:04:04.183 19:00:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:04.183 19:00:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:04:04.183 19:00:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:04:04.183 [2024-11-26 19:00:16.786441] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:04.183 [2024-11-26 19:00:16.786521] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3600733 ] 00:04:04.443 [2024-11-26 19:00:16.871102] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:04.444 [2024-11-26 19:00:16.914109] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 3600733 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 3600733 ']' 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 3600733 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3600733 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3600733' 00:04:09.728 killing process with pid 3600733 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 3600733 00:04:09.728 19:00:21 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 3600733 00:04:09.728 00:04:09.728 real 0m5.287s 00:04:09.728 user 0m5.091s 00:04:09.728 sys 0m0.247s 00:04:09.728 19:00:22 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:09.728 19:00:22 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:09.728 ************************************ 00:04:09.728 END TEST skip_rpc 00:04:09.728 ************************************ 00:04:09.728 19:00:22 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:04:09.729 19:00:22 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:09.729 19:00:22 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:09.729 19:00:22 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:09.729 ************************************ 00:04:09.729 START TEST skip_rpc_with_json 00:04:09.729 ************************************ 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=3601916 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 3601916 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 3601916 ']' 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:09.729 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:09.729 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:09.729 [2024-11-26 19:00:22.149677] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:09.729 [2024-11-26 19:00:22.149732] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3601916 ] 00:04:09.729 [2024-11-26 19:00:22.233639] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:09.729 [2024-11-26 19:00:22.275095] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:10.669 [2024-11-26 19:00:22.958641] nvmf_rpc.c:2706:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:04:10.669 request: 00:04:10.669 { 00:04:10.669 "trtype": "tcp", 00:04:10.669 "method": "nvmf_get_transports", 00:04:10.669 "req_id": 1 00:04:10.669 } 00:04:10.669 Got JSON-RPC error response 00:04:10.669 response: 00:04:10.669 { 00:04:10.669 "code": -19, 00:04:10.669 "message": "No such device" 00:04:10.669 } 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:10.669 [2024-11-26 19:00:22.970755] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:10.669 19:00:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:10.669 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:10.669 19:00:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:10.669 { 00:04:10.669 "subsystems": [ 00:04:10.669 { 00:04:10.669 "subsystem": "fsdev", 00:04:10.669 "config": [ 00:04:10.669 { 00:04:10.669 "method": "fsdev_set_opts", 00:04:10.669 "params": { 00:04:10.669 "fsdev_io_pool_size": 65535, 00:04:10.669 "fsdev_io_cache_size": 256 00:04:10.669 } 00:04:10.669 } 00:04:10.669 ] 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "subsystem": "vfio_user_target", 00:04:10.669 "config": null 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "subsystem": "keyring", 00:04:10.669 "config": [] 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "subsystem": "iobuf", 00:04:10.669 "config": [ 00:04:10.669 { 00:04:10.669 "method": "iobuf_set_options", 00:04:10.669 "params": { 00:04:10.669 "small_pool_count": 8192, 00:04:10.669 "large_pool_count": 1024, 00:04:10.669 "small_bufsize": 8192, 00:04:10.669 "large_bufsize": 135168, 00:04:10.669 "enable_numa": false 00:04:10.669 } 00:04:10.669 } 00:04:10.669 ] 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "subsystem": "sock", 00:04:10.669 "config": [ 00:04:10.669 { 00:04:10.669 "method": "sock_set_default_impl", 00:04:10.669 "params": { 00:04:10.669 "impl_name": "posix" 00:04:10.669 } 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "method": "sock_impl_set_options", 00:04:10.669 "params": { 00:04:10.669 "impl_name": "ssl", 00:04:10.669 "recv_buf_size": 4096, 00:04:10.669 "send_buf_size": 4096, 00:04:10.669 "enable_recv_pipe": true, 00:04:10.669 "enable_quickack": false, 00:04:10.669 "enable_placement_id": 0, 00:04:10.669 "enable_zerocopy_send_server": true, 00:04:10.669 "enable_zerocopy_send_client": false, 00:04:10.669 "zerocopy_threshold": 0, 00:04:10.669 "tls_version": 0, 00:04:10.669 "enable_ktls": false 00:04:10.669 } 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "method": "sock_impl_set_options", 00:04:10.669 "params": { 00:04:10.669 "impl_name": "posix", 00:04:10.669 "recv_buf_size": 2097152, 00:04:10.669 "send_buf_size": 2097152, 00:04:10.669 "enable_recv_pipe": true, 00:04:10.669 "enable_quickack": false, 00:04:10.669 "enable_placement_id": 0, 00:04:10.669 "enable_zerocopy_send_server": true, 00:04:10.669 "enable_zerocopy_send_client": false, 00:04:10.669 "zerocopy_threshold": 0, 00:04:10.669 "tls_version": 0, 00:04:10.669 "enable_ktls": false 00:04:10.669 } 00:04:10.669 } 00:04:10.669 ] 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "subsystem": "vmd", 00:04:10.669 "config": [] 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "subsystem": "accel", 00:04:10.669 "config": [ 00:04:10.669 { 00:04:10.669 "method": "accel_set_options", 00:04:10.669 "params": { 00:04:10.669 "small_cache_size": 128, 00:04:10.669 "large_cache_size": 16, 00:04:10.669 "task_count": 2048, 00:04:10.669 "sequence_count": 2048, 00:04:10.669 "buf_count": 2048 00:04:10.669 } 00:04:10.669 } 00:04:10.669 ] 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "subsystem": "bdev", 00:04:10.669 "config": [ 00:04:10.669 { 00:04:10.669 "method": "bdev_set_options", 00:04:10.669 "params": { 00:04:10.669 "bdev_io_pool_size": 65535, 00:04:10.669 "bdev_io_cache_size": 256, 00:04:10.669 "bdev_auto_examine": true, 00:04:10.669 "iobuf_small_cache_size": 128, 00:04:10.669 "iobuf_large_cache_size": 16 00:04:10.669 } 00:04:10.669 }, 00:04:10.669 { 00:04:10.669 "method": "bdev_raid_set_options", 00:04:10.669 "params": { 00:04:10.669 "process_window_size_kb": 1024, 00:04:10.669 "process_max_bandwidth_mb_sec": 0 00:04:10.669 } 00:04:10.669 }, 00:04:10.669 { 00:04:10.670 "method": "bdev_iscsi_set_options", 00:04:10.670 "params": { 00:04:10.670 "timeout_sec": 30 00:04:10.670 } 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "method": "bdev_nvme_set_options", 00:04:10.670 "params": { 00:04:10.670 "action_on_timeout": "none", 00:04:10.670 "timeout_us": 0, 00:04:10.670 "timeout_admin_us": 0, 00:04:10.670 "keep_alive_timeout_ms": 10000, 00:04:10.670 "arbitration_burst": 0, 00:04:10.670 "low_priority_weight": 0, 00:04:10.670 "medium_priority_weight": 0, 00:04:10.670 "high_priority_weight": 0, 00:04:10.670 "nvme_adminq_poll_period_us": 10000, 00:04:10.670 "nvme_ioq_poll_period_us": 0, 00:04:10.670 "io_queue_requests": 0, 00:04:10.670 "delay_cmd_submit": true, 00:04:10.670 "transport_retry_count": 4, 00:04:10.670 "bdev_retry_count": 3, 00:04:10.670 "transport_ack_timeout": 0, 00:04:10.670 "ctrlr_loss_timeout_sec": 0, 00:04:10.670 "reconnect_delay_sec": 0, 00:04:10.670 "fast_io_fail_timeout_sec": 0, 00:04:10.670 "disable_auto_failback": false, 00:04:10.670 "generate_uuids": false, 00:04:10.670 "transport_tos": 0, 00:04:10.670 "nvme_error_stat": false, 00:04:10.670 "rdma_srq_size": 0, 00:04:10.670 "io_path_stat": false, 00:04:10.670 "allow_accel_sequence": false, 00:04:10.670 "rdma_max_cq_size": 0, 00:04:10.670 "rdma_cm_event_timeout_ms": 0, 00:04:10.670 "dhchap_digests": [ 00:04:10.670 "sha256", 00:04:10.670 "sha384", 00:04:10.670 "sha512" 00:04:10.670 ], 00:04:10.670 "dhchap_dhgroups": [ 00:04:10.670 "null", 00:04:10.670 "ffdhe2048", 00:04:10.670 "ffdhe3072", 00:04:10.670 "ffdhe4096", 00:04:10.670 "ffdhe6144", 00:04:10.670 "ffdhe8192" 00:04:10.670 ] 00:04:10.670 } 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "method": "bdev_nvme_set_hotplug", 00:04:10.670 "params": { 00:04:10.670 "period_us": 100000, 00:04:10.670 "enable": false 00:04:10.670 } 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "method": "bdev_wait_for_examine" 00:04:10.670 } 00:04:10.670 ] 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "scsi", 00:04:10.670 "config": null 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "scheduler", 00:04:10.670 "config": [ 00:04:10.670 { 00:04:10.670 "method": "framework_set_scheduler", 00:04:10.670 "params": { 00:04:10.670 "name": "static" 00:04:10.670 } 00:04:10.670 } 00:04:10.670 ] 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "vhost_scsi", 00:04:10.670 "config": [] 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "vhost_blk", 00:04:10.670 "config": [] 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "ublk", 00:04:10.670 "config": [] 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "nbd", 00:04:10.670 "config": [] 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "nvmf", 00:04:10.670 "config": [ 00:04:10.670 { 00:04:10.670 "method": "nvmf_set_config", 00:04:10.670 "params": { 00:04:10.670 "discovery_filter": "match_any", 00:04:10.670 "admin_cmd_passthru": { 00:04:10.670 "identify_ctrlr": false 00:04:10.670 }, 00:04:10.670 "dhchap_digests": [ 00:04:10.670 "sha256", 00:04:10.670 "sha384", 00:04:10.670 "sha512" 00:04:10.670 ], 00:04:10.670 "dhchap_dhgroups": [ 00:04:10.670 "null", 00:04:10.670 "ffdhe2048", 00:04:10.670 "ffdhe3072", 00:04:10.670 "ffdhe4096", 00:04:10.670 "ffdhe6144", 00:04:10.670 "ffdhe8192" 00:04:10.670 ] 00:04:10.670 } 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "method": "nvmf_set_max_subsystems", 00:04:10.670 "params": { 00:04:10.670 "max_subsystems": 1024 00:04:10.670 } 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "method": "nvmf_set_crdt", 00:04:10.670 "params": { 00:04:10.670 "crdt1": 0, 00:04:10.670 "crdt2": 0, 00:04:10.670 "crdt3": 0 00:04:10.670 } 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "method": "nvmf_create_transport", 00:04:10.670 "params": { 00:04:10.670 "trtype": "TCP", 00:04:10.670 "max_queue_depth": 128, 00:04:10.670 "max_io_qpairs_per_ctrlr": 127, 00:04:10.670 "in_capsule_data_size": 4096, 00:04:10.670 "max_io_size": 131072, 00:04:10.670 "io_unit_size": 131072, 00:04:10.670 "max_aq_depth": 128, 00:04:10.670 "num_shared_buffers": 511, 00:04:10.670 "buf_cache_size": 4294967295, 00:04:10.670 "dif_insert_or_strip": false, 00:04:10.670 "zcopy": false, 00:04:10.670 "c2h_success": true, 00:04:10.670 "sock_priority": 0, 00:04:10.670 "abort_timeout_sec": 1, 00:04:10.670 "ack_timeout": 0, 00:04:10.670 "data_wr_pool_size": 0 00:04:10.670 } 00:04:10.670 } 00:04:10.670 ] 00:04:10.670 }, 00:04:10.670 { 00:04:10.670 "subsystem": "iscsi", 00:04:10.670 "config": [ 00:04:10.670 { 00:04:10.670 "method": "iscsi_set_options", 00:04:10.670 "params": { 00:04:10.670 "node_base": "iqn.2016-06.io.spdk", 00:04:10.670 "max_sessions": 128, 00:04:10.670 "max_connections_per_session": 2, 00:04:10.670 "max_queue_depth": 64, 00:04:10.670 "default_time2wait": 2, 00:04:10.670 "default_time2retain": 20, 00:04:10.670 "first_burst_length": 8192, 00:04:10.670 "immediate_data": true, 00:04:10.670 "allow_duplicated_isid": false, 00:04:10.670 "error_recovery_level": 0, 00:04:10.670 "nop_timeout": 60, 00:04:10.670 "nop_in_interval": 30, 00:04:10.670 "disable_chap": false, 00:04:10.670 "require_chap": false, 00:04:10.670 "mutual_chap": false, 00:04:10.670 "chap_group": 0, 00:04:10.670 "max_large_datain_per_connection": 64, 00:04:10.670 "max_r2t_per_connection": 4, 00:04:10.670 "pdu_pool_size": 36864, 00:04:10.670 "immediate_data_pool_size": 16384, 00:04:10.670 "data_out_pool_size": 2048 00:04:10.670 } 00:04:10.670 } 00:04:10.670 ] 00:04:10.670 } 00:04:10.670 ] 00:04:10.670 } 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 3601916 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 3601916 ']' 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 3601916 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3601916 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3601916' 00:04:10.670 killing process with pid 3601916 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 3601916 00:04:10.670 19:00:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 3601916 00:04:10.929 19:00:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=3602115 00:04:10.929 19:00:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:04:10.929 19:00:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 3602115 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 3602115 ']' 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 3602115 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3602115 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3602115' 00:04:16.215 killing process with pid 3602115 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 3602115 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 3602115 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:04:16.215 00:04:16.215 real 0m6.617s 00:04:16.215 user 0m6.506s 00:04:16.215 sys 0m0.580s 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:16.215 ************************************ 00:04:16.215 END TEST skip_rpc_with_json 00:04:16.215 ************************************ 00:04:16.215 19:00:28 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:04:16.215 19:00:28 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:16.215 19:00:28 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:16.215 19:00:28 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:16.215 ************************************ 00:04:16.215 START TEST skip_rpc_with_delay 00:04:16.215 ************************************ 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:04:16.215 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:04:16.477 [2024-11-26 19:00:28.851013] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:04:16.477 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:04:16.477 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:04:16.477 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:04:16.477 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:04:16.477 00:04:16.477 real 0m0.080s 00:04:16.477 user 0m0.054s 00:04:16.477 sys 0m0.025s 00:04:16.477 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:16.477 19:00:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:04:16.477 ************************************ 00:04:16.477 END TEST skip_rpc_with_delay 00:04:16.477 ************************************ 00:04:16.477 19:00:28 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:04:16.477 19:00:28 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:04:16.477 19:00:28 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:04:16.477 19:00:28 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:16.477 19:00:28 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:16.477 19:00:28 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:16.477 ************************************ 00:04:16.477 START TEST exit_on_failed_rpc_init 00:04:16.477 ************************************ 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=3603420 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 3603420 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 3603420 ']' 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:16.477 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:16.477 19:00:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:04:16.477 [2024-11-26 19:00:29.010183] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:16.477 [2024-11-26 19:00:29.010235] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3603420 ] 00:04:16.477 [2024-11-26 19:00:29.091274] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:16.737 [2024-11-26 19:00:29.128820] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:17.306 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:17.307 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:17.307 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:04:17.307 19:00:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:04:17.307 [2024-11-26 19:00:29.873090] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:17.307 [2024-11-26 19:00:29.873145] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3603514 ] 00:04:17.567 [2024-11-26 19:00:29.966792] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:17.567 [2024-11-26 19:00:30.003083] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:17.567 [2024-11-26 19:00:30.003140] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:04:17.567 [2024-11-26 19:00:30.003150] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:04:17.567 [2024-11-26 19:00:30.003158] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 3603420 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 3603420 ']' 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 3603420 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3603420 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3603420' 00:04:17.567 killing process with pid 3603420 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 3603420 00:04:17.567 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 3603420 00:04:17.829 00:04:17.829 real 0m1.366s 00:04:17.829 user 0m1.600s 00:04:17.829 sys 0m0.396s 00:04:17.829 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:17.829 19:00:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:04:17.829 ************************************ 00:04:17.829 END TEST exit_on_failed_rpc_init 00:04:17.829 ************************************ 00:04:17.829 19:00:30 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:17.829 00:04:17.829 real 0m13.864s 00:04:17.829 user 0m13.485s 00:04:17.829 sys 0m1.559s 00:04:17.829 19:00:30 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:17.829 19:00:30 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:17.829 ************************************ 00:04:17.829 END TEST skip_rpc 00:04:17.829 ************************************ 00:04:17.829 19:00:30 -- spdk/autotest.sh@158 -- # run_test rpc_client /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:04:17.829 19:00:30 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:17.829 19:00:30 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:17.829 19:00:30 -- common/autotest_common.sh@10 -- # set +x 00:04:17.829 ************************************ 00:04:17.829 START TEST rpc_client 00:04:17.829 ************************************ 00:04:17.829 19:00:30 rpc_client -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:04:18.091 * Looking for test storage... 00:04:18.091 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1693 -- # lcov --version 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@345 -- # : 1 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@353 -- # local d=1 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@355 -- # echo 1 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@353 -- # local d=2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@355 -- # echo 2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:18.091 19:00:30 rpc_client -- scripts/common.sh@368 -- # return 0 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:18.091 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.091 --rc genhtml_branch_coverage=1 00:04:18.091 --rc genhtml_function_coverage=1 00:04:18.091 --rc genhtml_legend=1 00:04:18.091 --rc geninfo_all_blocks=1 00:04:18.091 --rc geninfo_unexecuted_blocks=1 00:04:18.091 00:04:18.091 ' 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:18.091 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.091 --rc genhtml_branch_coverage=1 00:04:18.091 --rc genhtml_function_coverage=1 00:04:18.091 --rc genhtml_legend=1 00:04:18.091 --rc geninfo_all_blocks=1 00:04:18.091 --rc geninfo_unexecuted_blocks=1 00:04:18.091 00:04:18.091 ' 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:18.091 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.091 --rc genhtml_branch_coverage=1 00:04:18.091 --rc genhtml_function_coverage=1 00:04:18.091 --rc genhtml_legend=1 00:04:18.091 --rc geninfo_all_blocks=1 00:04:18.091 --rc geninfo_unexecuted_blocks=1 00:04:18.091 00:04:18.091 ' 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:18.091 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.091 --rc genhtml_branch_coverage=1 00:04:18.091 --rc genhtml_function_coverage=1 00:04:18.091 --rc genhtml_legend=1 00:04:18.091 --rc geninfo_all_blocks=1 00:04:18.091 --rc geninfo_unexecuted_blocks=1 00:04:18.091 00:04:18.091 ' 00:04:18.091 19:00:30 rpc_client -- rpc_client/rpc_client.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client_test 00:04:18.091 OK 00:04:18.091 19:00:30 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:04:18.091 00:04:18.091 real 0m0.232s 00:04:18.091 user 0m0.141s 00:04:18.091 sys 0m0.104s 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:18.091 19:00:30 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:04:18.091 ************************************ 00:04:18.091 END TEST rpc_client 00:04:18.091 ************************************ 00:04:18.091 19:00:30 -- spdk/autotest.sh@159 -- # run_test json_config /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:04:18.091 19:00:30 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:18.091 19:00:30 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:18.091 19:00:30 -- common/autotest_common.sh@10 -- # set +x 00:04:18.354 ************************************ 00:04:18.354 START TEST json_config 00:04:18.354 ************************************ 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1693 -- # lcov --version 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:18.354 19:00:30 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:18.354 19:00:30 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:04:18.354 19:00:30 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:04:18.354 19:00:30 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:04:18.354 19:00:30 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:04:18.354 19:00:30 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:04:18.354 19:00:30 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:18.354 19:00:30 json_config -- scripts/common.sh@344 -- # case "$op" in 00:04:18.354 19:00:30 json_config -- scripts/common.sh@345 -- # : 1 00:04:18.354 19:00:30 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:18.354 19:00:30 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:18.354 19:00:30 json_config -- scripts/common.sh@365 -- # decimal 1 00:04:18.354 19:00:30 json_config -- scripts/common.sh@353 -- # local d=1 00:04:18.354 19:00:30 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:18.354 19:00:30 json_config -- scripts/common.sh@355 -- # echo 1 00:04:18.354 19:00:30 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:04:18.354 19:00:30 json_config -- scripts/common.sh@366 -- # decimal 2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@353 -- # local d=2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:18.354 19:00:30 json_config -- scripts/common.sh@355 -- # echo 2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:04:18.354 19:00:30 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:18.354 19:00:30 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:18.354 19:00:30 json_config -- scripts/common.sh@368 -- # return 0 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:18.354 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.354 --rc genhtml_branch_coverage=1 00:04:18.354 --rc genhtml_function_coverage=1 00:04:18.354 --rc genhtml_legend=1 00:04:18.354 --rc geninfo_all_blocks=1 00:04:18.354 --rc geninfo_unexecuted_blocks=1 00:04:18.354 00:04:18.354 ' 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:18.354 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.354 --rc genhtml_branch_coverage=1 00:04:18.354 --rc genhtml_function_coverage=1 00:04:18.354 --rc genhtml_legend=1 00:04:18.354 --rc geninfo_all_blocks=1 00:04:18.354 --rc geninfo_unexecuted_blocks=1 00:04:18.354 00:04:18.354 ' 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:18.354 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.354 --rc genhtml_branch_coverage=1 00:04:18.354 --rc genhtml_function_coverage=1 00:04:18.354 --rc genhtml_legend=1 00:04:18.354 --rc geninfo_all_blocks=1 00:04:18.354 --rc geninfo_unexecuted_blocks=1 00:04:18.354 00:04:18.354 ' 00:04:18.354 19:00:30 json_config -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:18.354 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:18.354 --rc genhtml_branch_coverage=1 00:04:18.354 --rc genhtml_function_coverage=1 00:04:18.354 --rc genhtml_legend=1 00:04:18.354 --rc geninfo_all_blocks=1 00:04:18.354 --rc geninfo_unexecuted_blocks=1 00:04:18.354 00:04:18.354 ' 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@7 -- # uname -s 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:18.354 19:00:30 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:04:18.354 19:00:30 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:18.354 19:00:30 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:18.354 19:00:30 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:18.354 19:00:30 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:18.354 19:00:30 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:18.354 19:00:30 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:18.354 19:00:30 json_config -- paths/export.sh@5 -- # export PATH 00:04:18.354 19:00:30 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@51 -- # : 0 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:18.354 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:18.354 19:00:30 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@31 -- # app_pid=(['target']='' ['initiator']='') 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@31 -- # declare -A app_pid 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@32 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock' ['initiator']='/var/tmp/spdk_initiator.sock') 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@32 -- # declare -A app_socket 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@33 -- # app_params=(['target']='-m 0x1 -s 1024' ['initiator']='-m 0x2 -g -u -s 1024') 00:04:18.354 19:00:30 json_config -- json_config/json_config.sh@33 -- # declare -A app_params 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@34 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json' ['initiator']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json') 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@34 -- # declare -A configs_path 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@40 -- # last_event_id=0 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@362 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@363 -- # echo 'INFO: JSON configuration test init' 00:04:18.355 INFO: JSON configuration test init 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@364 -- # json_config_test_init 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@269 -- # timing_enter json_config_test_init 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@270 -- # timing_enter json_config_setup_target 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:18.355 19:00:30 json_config -- json_config/json_config.sh@272 -- # json_config_test_start_app target --wait-for-rpc 00:04:18.355 19:00:30 json_config -- json_config/common.sh@9 -- # local app=target 00:04:18.355 19:00:30 json_config -- json_config/common.sh@10 -- # shift 00:04:18.355 19:00:30 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:04:18.355 19:00:30 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:04:18.355 19:00:30 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:04:18.355 19:00:30 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:18.355 19:00:30 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:18.355 19:00:30 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=3603965 00:04:18.355 19:00:30 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:04:18.355 Waiting for target to run... 00:04:18.355 19:00:30 json_config -- json_config/common.sh@25 -- # waitforlisten 3603965 /var/tmp/spdk_tgt.sock 00:04:18.355 19:00:30 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --wait-for-rpc 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@835 -- # '[' -z 3603965 ']' 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:04:18.355 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:18.355 19:00:30 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:18.615 [2024-11-26 19:00:31.026246] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:18.615 [2024-11-26 19:00:31.026310] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3603965 ] 00:04:18.876 [2024-11-26 19:00:31.283766] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:18.876 [2024-11-26 19:00:31.311895] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:19.449 19:00:31 json_config -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:19.449 19:00:31 json_config -- common/autotest_common.sh@868 -- # return 0 00:04:19.449 19:00:31 json_config -- json_config/common.sh@26 -- # echo '' 00:04:19.449 00:04:19.449 19:00:31 json_config -- json_config/json_config.sh@276 -- # create_accel_config 00:04:19.449 19:00:31 json_config -- json_config/json_config.sh@100 -- # timing_enter create_accel_config 00:04:19.449 19:00:31 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:19.449 19:00:31 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:19.449 19:00:31 json_config -- json_config/json_config.sh@102 -- # [[ 0 -eq 1 ]] 00:04:19.449 19:00:31 json_config -- json_config/json_config.sh@108 -- # timing_exit create_accel_config 00:04:19.449 19:00:31 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:19.449 19:00:31 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:19.449 19:00:31 json_config -- json_config/json_config.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh --json-with-subsystems 00:04:19.449 19:00:31 json_config -- json_config/json_config.sh@281 -- # tgt_rpc load_config 00:04:19.449 19:00:31 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock load_config 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@283 -- # tgt_check_notification_types 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@43 -- # timing_enter tgt_check_notification_types 00:04:20.022 19:00:32 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:20.022 19:00:32 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@45 -- # local ret=0 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@46 -- # enabled_types=('bdev_register' 'bdev_unregister') 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@46 -- # local enabled_types 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@47 -- # [[ y == y ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@48 -- # enabled_types+=("fsdev_register" "fsdev_unregister") 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@51 -- # jq -r '.[]' 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@51 -- # tgt_rpc notify_get_types 00:04:20.022 19:00:32 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_types 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@51 -- # get_types=('fsdev_register' 'fsdev_unregister' 'bdev_register' 'bdev_unregister') 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@51 -- # local get_types 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@53 -- # local type_diff 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@54 -- # echo bdev_register bdev_unregister fsdev_register fsdev_unregister fsdev_register fsdev_unregister bdev_register bdev_unregister 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@54 -- # tr ' ' '\n' 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@54 -- # sort 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@54 -- # uniq -u 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@54 -- # type_diff= 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@56 -- # [[ -n '' ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@61 -- # timing_exit tgt_check_notification_types 00:04:20.022 19:00:32 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:20.022 19:00:32 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@62 -- # return 0 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@285 -- # [[ 0 -eq 1 ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@289 -- # [[ 0 -eq 1 ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@293 -- # [[ 0 -eq 1 ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@297 -- # [[ 1 -eq 1 ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@298 -- # create_nvmf_subsystem_config 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@237 -- # timing_enter create_nvmf_subsystem_config 00:04:20.022 19:00:32 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:20.022 19:00:32 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@239 -- # NVMF_FIRST_TARGET_IP=127.0.0.1 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@240 -- # [[ tcp == \r\d\m\a ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@244 -- # [[ -z 127.0.0.1 ]] 00:04:20.022 19:00:32 json_config -- json_config/json_config.sh@249 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocForNvmf0 00:04:20.022 19:00:32 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocForNvmf0 00:04:20.283 MallocForNvmf0 00:04:20.283 19:00:32 json_config -- json_config/json_config.sh@250 -- # tgt_rpc bdev_malloc_create 4 1024 --name MallocForNvmf1 00:04:20.283 19:00:32 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 4 1024 --name MallocForNvmf1 00:04:20.543 MallocForNvmf1 00:04:20.543 19:00:33 json_config -- json_config/json_config.sh@252 -- # tgt_rpc nvmf_create_transport -t tcp -u 8192 -c 0 00:04:20.543 19:00:33 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_transport -t tcp -u 8192 -c 0 00:04:20.543 [2024-11-26 19:00:33.153683] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:20.803 19:00:33 json_config -- json_config/json_config.sh@253 -- # tgt_rpc nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:04:20.803 19:00:33 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:04:20.803 19:00:33 json_config -- json_config/json_config.sh@254 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:04:20.803 19:00:33 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:04:21.062 19:00:33 json_config -- json_config/json_config.sh@255 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:04:21.062 19:00:33 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:04:21.321 19:00:33 json_config -- json_config/json_config.sh@256 -- # tgt_rpc nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:04:21.322 19:00:33 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:04:21.322 [2024-11-26 19:00:33.859953] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:04:21.322 19:00:33 json_config -- json_config/json_config.sh@258 -- # timing_exit create_nvmf_subsystem_config 00:04:21.322 19:00:33 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:21.322 19:00:33 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:21.322 19:00:33 json_config -- json_config/json_config.sh@300 -- # timing_exit json_config_setup_target 00:04:21.322 19:00:33 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:21.322 19:00:33 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:21.581 19:00:33 json_config -- json_config/json_config.sh@302 -- # [[ 0 -eq 1 ]] 00:04:21.581 19:00:33 json_config -- json_config/json_config.sh@307 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:04:21.581 19:00:33 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:04:21.581 MallocBdevForConfigChangeCheck 00:04:21.581 19:00:34 json_config -- json_config/json_config.sh@309 -- # timing_exit json_config_test_init 00:04:21.581 19:00:34 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:21.581 19:00:34 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:21.581 19:00:34 json_config -- json_config/json_config.sh@366 -- # tgt_rpc save_config 00:04:21.581 19:00:34 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:22.149 19:00:34 json_config -- json_config/json_config.sh@368 -- # echo 'INFO: shutting down applications...' 00:04:22.149 INFO: shutting down applications... 00:04:22.149 19:00:34 json_config -- json_config/json_config.sh@369 -- # [[ 0 -eq 1 ]] 00:04:22.149 19:00:34 json_config -- json_config/json_config.sh@375 -- # json_config_clear target 00:04:22.149 19:00:34 json_config -- json_config/json_config.sh@339 -- # [[ -n 22 ]] 00:04:22.149 19:00:34 json_config -- json_config/json_config.sh@340 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py -s /var/tmp/spdk_tgt.sock clear_config 00:04:22.409 Calling clear_iscsi_subsystem 00:04:22.409 Calling clear_nvmf_subsystem 00:04:22.409 Calling clear_nbd_subsystem 00:04:22.409 Calling clear_ublk_subsystem 00:04:22.409 Calling clear_vhost_blk_subsystem 00:04:22.409 Calling clear_vhost_scsi_subsystem 00:04:22.409 Calling clear_bdev_subsystem 00:04:22.409 19:00:34 json_config -- json_config/json_config.sh@344 -- # local config_filter=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py 00:04:22.409 19:00:34 json_config -- json_config/json_config.sh@350 -- # count=100 00:04:22.409 19:00:34 json_config -- json_config/json_config.sh@351 -- # '[' 100 -gt 0 ']' 00:04:22.409 19:00:34 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:22.409 19:00:34 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method delete_global_parameters 00:04:22.409 19:00:34 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method check_empty 00:04:22.667 19:00:35 json_config -- json_config/json_config.sh@352 -- # break 00:04:22.667 19:00:35 json_config -- json_config/json_config.sh@357 -- # '[' 100 -eq 0 ']' 00:04:22.667 19:00:35 json_config -- json_config/json_config.sh@376 -- # json_config_test_shutdown_app target 00:04:22.667 19:00:35 json_config -- json_config/common.sh@31 -- # local app=target 00:04:22.667 19:00:35 json_config -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:04:22.667 19:00:35 json_config -- json_config/common.sh@35 -- # [[ -n 3603965 ]] 00:04:22.667 19:00:35 json_config -- json_config/common.sh@38 -- # kill -SIGINT 3603965 00:04:22.667 19:00:35 json_config -- json_config/common.sh@40 -- # (( i = 0 )) 00:04:22.667 19:00:35 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:22.667 19:00:35 json_config -- json_config/common.sh@41 -- # kill -0 3603965 00:04:22.667 19:00:35 json_config -- json_config/common.sh@45 -- # sleep 0.5 00:04:23.234 19:00:35 json_config -- json_config/common.sh@40 -- # (( i++ )) 00:04:23.234 19:00:35 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:23.234 19:00:35 json_config -- json_config/common.sh@41 -- # kill -0 3603965 00:04:23.234 19:00:35 json_config -- json_config/common.sh@42 -- # app_pid["$app"]= 00:04:23.234 19:00:35 json_config -- json_config/common.sh@43 -- # break 00:04:23.234 19:00:35 json_config -- json_config/common.sh@48 -- # [[ -n '' ]] 00:04:23.234 19:00:35 json_config -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:04:23.234 SPDK target shutdown done 00:04:23.234 19:00:35 json_config -- json_config/json_config.sh@378 -- # echo 'INFO: relaunching applications...' 00:04:23.234 INFO: relaunching applications... 00:04:23.234 19:00:35 json_config -- json_config/json_config.sh@379 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:23.234 19:00:35 json_config -- json_config/common.sh@9 -- # local app=target 00:04:23.234 19:00:35 json_config -- json_config/common.sh@10 -- # shift 00:04:23.234 19:00:35 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:04:23.234 19:00:35 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:04:23.234 19:00:35 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:04:23.234 19:00:35 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:23.234 19:00:35 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:23.234 19:00:35 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=3605095 00:04:23.234 19:00:35 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:04:23.234 Waiting for target to run... 00:04:23.234 19:00:35 json_config -- json_config/common.sh@25 -- # waitforlisten 3605095 /var/tmp/spdk_tgt.sock 00:04:23.234 19:00:35 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:23.234 19:00:35 json_config -- common/autotest_common.sh@835 -- # '[' -z 3605095 ']' 00:04:23.234 19:00:35 json_config -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:04:23.234 19:00:35 json_config -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:23.234 19:00:35 json_config -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:04:23.234 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:04:23.234 19:00:35 json_config -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:23.234 19:00:35 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:23.234 [2024-11-26 19:00:35.801675] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:23.234 [2024-11-26 19:00:35.801743] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3605095 ] 00:04:23.804 [2024-11-26 19:00:36.118107] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:23.804 [2024-11-26 19:00:36.150598] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:24.064 [2024-11-26 19:00:36.676310] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:24.324 [2024-11-26 19:00:36.708690] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:04:24.324 19:00:36 json_config -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:24.324 19:00:36 json_config -- common/autotest_common.sh@868 -- # return 0 00:04:24.324 19:00:36 json_config -- json_config/common.sh@26 -- # echo '' 00:04:24.324 00:04:24.324 19:00:36 json_config -- json_config/json_config.sh@380 -- # [[ 0 -eq 1 ]] 00:04:24.324 19:00:36 json_config -- json_config/json_config.sh@384 -- # echo 'INFO: Checking if target configuration is the same...' 00:04:24.324 INFO: Checking if target configuration is the same... 00:04:24.324 19:00:36 json_config -- json_config/json_config.sh@385 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:24.324 19:00:36 json_config -- json_config/json_config.sh@385 -- # tgt_rpc save_config 00:04:24.325 19:00:36 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:24.325 + '[' 2 -ne 2 ']' 00:04:24.325 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:04:24.325 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:04:24.325 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:24.325 +++ basename /dev/fd/62 00:04:24.325 ++ mktemp /tmp/62.XXX 00:04:24.325 + tmp_file_1=/tmp/62.ohm 00:04:24.325 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:24.325 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:04:24.325 + tmp_file_2=/tmp/spdk_tgt_config.json.d8Y 00:04:24.325 + ret=0 00:04:24.325 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:24.584 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:24.584 + diff -u /tmp/62.ohm /tmp/spdk_tgt_config.json.d8Y 00:04:24.584 + echo 'INFO: JSON config files are the same' 00:04:24.584 INFO: JSON config files are the same 00:04:24.584 + rm /tmp/62.ohm /tmp/spdk_tgt_config.json.d8Y 00:04:24.584 + exit 0 00:04:24.584 19:00:37 json_config -- json_config/json_config.sh@386 -- # [[ 0 -eq 1 ]] 00:04:24.584 19:00:37 json_config -- json_config/json_config.sh@391 -- # echo 'INFO: changing configuration and checking if this can be detected...' 00:04:24.584 INFO: changing configuration and checking if this can be detected... 00:04:24.584 19:00:37 json_config -- json_config/json_config.sh@393 -- # tgt_rpc bdev_malloc_delete MallocBdevForConfigChangeCheck 00:04:24.584 19:00:37 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_delete MallocBdevForConfigChangeCheck 00:04:24.845 19:00:37 json_config -- json_config/json_config.sh@394 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:24.845 19:00:37 json_config -- json_config/json_config.sh@394 -- # tgt_rpc save_config 00:04:24.845 19:00:37 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:24.845 + '[' 2 -ne 2 ']' 00:04:24.845 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:04:24.845 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:04:24.845 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:24.845 +++ basename /dev/fd/62 00:04:24.845 ++ mktemp /tmp/62.XXX 00:04:24.845 + tmp_file_1=/tmp/62.vvC 00:04:24.845 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:24.845 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:04:24.845 + tmp_file_2=/tmp/spdk_tgt_config.json.Oww 00:04:24.845 + ret=0 00:04:24.845 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:25.106 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:25.106 + diff -u /tmp/62.vvC /tmp/spdk_tgt_config.json.Oww 00:04:25.106 + ret=1 00:04:25.106 + echo '=== Start of file: /tmp/62.vvC ===' 00:04:25.106 + cat /tmp/62.vvC 00:04:25.106 + echo '=== End of file: /tmp/62.vvC ===' 00:04:25.106 + echo '' 00:04:25.106 + echo '=== Start of file: /tmp/spdk_tgt_config.json.Oww ===' 00:04:25.106 + cat /tmp/spdk_tgt_config.json.Oww 00:04:25.106 + echo '=== End of file: /tmp/spdk_tgt_config.json.Oww ===' 00:04:25.106 + echo '' 00:04:25.106 + rm /tmp/62.vvC /tmp/spdk_tgt_config.json.Oww 00:04:25.106 + exit 1 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@398 -- # echo 'INFO: configuration change detected.' 00:04:25.106 INFO: configuration change detected. 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@401 -- # json_config_test_fini 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@313 -- # timing_enter json_config_test_fini 00:04:25.106 19:00:37 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:25.106 19:00:37 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@314 -- # local ret=0 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@316 -- # [[ -n '' ]] 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@324 -- # [[ -n 3605095 ]] 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@327 -- # cleanup_bdev_subsystem_config 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@191 -- # timing_enter cleanup_bdev_subsystem_config 00:04:25.106 19:00:37 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:25.106 19:00:37 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@193 -- # [[ 0 -eq 1 ]] 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@200 -- # uname -s 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@200 -- # [[ Linux = Linux ]] 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@201 -- # rm -f /sample_aio 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@204 -- # [[ 0 -eq 1 ]] 00:04:25.106 19:00:37 json_config -- json_config/json_config.sh@208 -- # timing_exit cleanup_bdev_subsystem_config 00:04:25.106 19:00:37 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:25.106 19:00:37 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:25.368 19:00:37 json_config -- json_config/json_config.sh@330 -- # killprocess 3605095 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@954 -- # '[' -z 3605095 ']' 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@958 -- # kill -0 3605095 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@959 -- # uname 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3605095 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3605095' 00:04:25.368 killing process with pid 3605095 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@973 -- # kill 3605095 00:04:25.368 19:00:37 json_config -- common/autotest_common.sh@978 -- # wait 3605095 00:04:25.628 19:00:38 json_config -- json_config/json_config.sh@333 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:25.628 19:00:38 json_config -- json_config/json_config.sh@334 -- # timing_exit json_config_test_fini 00:04:25.628 19:00:38 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:25.628 19:00:38 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:25.628 19:00:38 json_config -- json_config/json_config.sh@335 -- # return 0 00:04:25.628 19:00:38 json_config -- json_config/json_config.sh@403 -- # echo 'INFO: Success' 00:04:25.628 INFO: Success 00:04:25.628 00:04:25.628 real 0m7.390s 00:04:25.628 user 0m8.903s 00:04:25.628 sys 0m1.945s 00:04:25.628 19:00:38 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:25.628 19:00:38 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:25.628 ************************************ 00:04:25.628 END TEST json_config 00:04:25.628 ************************************ 00:04:25.628 19:00:38 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:04:25.628 19:00:38 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:25.628 19:00:38 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:25.628 19:00:38 -- common/autotest_common.sh@10 -- # set +x 00:04:25.628 ************************************ 00:04:25.628 START TEST json_config_extra_key 00:04:25.628 ************************************ 00:04:25.628 19:00:38 json_config_extra_key -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1693 -- # lcov --version 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:25.889 19:00:38 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:25.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:25.889 --rc genhtml_branch_coverage=1 00:04:25.889 --rc genhtml_function_coverage=1 00:04:25.889 --rc genhtml_legend=1 00:04:25.889 --rc geninfo_all_blocks=1 00:04:25.889 --rc geninfo_unexecuted_blocks=1 00:04:25.889 00:04:25.889 ' 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:25.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:25.889 --rc genhtml_branch_coverage=1 00:04:25.889 --rc genhtml_function_coverage=1 00:04:25.889 --rc genhtml_legend=1 00:04:25.889 --rc geninfo_all_blocks=1 00:04:25.889 --rc geninfo_unexecuted_blocks=1 00:04:25.889 00:04:25.889 ' 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:25.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:25.889 --rc genhtml_branch_coverage=1 00:04:25.889 --rc genhtml_function_coverage=1 00:04:25.889 --rc genhtml_legend=1 00:04:25.889 --rc geninfo_all_blocks=1 00:04:25.889 --rc geninfo_unexecuted_blocks=1 00:04:25.889 00:04:25.889 ' 00:04:25.889 19:00:38 json_config_extra_key -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:25.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:25.889 --rc genhtml_branch_coverage=1 00:04:25.890 --rc genhtml_function_coverage=1 00:04:25.890 --rc genhtml_legend=1 00:04:25.890 --rc geninfo_all_blocks=1 00:04:25.890 --rc geninfo_unexecuted_blocks=1 00:04:25.890 00:04:25.890 ' 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:25.890 19:00:38 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:04:25.890 19:00:38 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:25.890 19:00:38 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:25.890 19:00:38 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:25.890 19:00:38 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.890 19:00:38 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.890 19:00:38 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.890 19:00:38 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:04:25.890 19:00:38 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:25.890 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:25.890 19:00:38 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json') 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:04:25.890 INFO: launching applications... 00:04:25.890 19:00:38 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=3605573 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:04:25.890 Waiting for target to run... 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 3605573 /var/tmp/spdk_tgt.sock 00:04:25.890 19:00:38 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 3605573 ']' 00:04:25.890 19:00:38 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:04:25.890 19:00:38 json_config_extra_key -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:04:25.890 19:00:38 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:25.890 19:00:38 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:04:25.890 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:04:25.890 19:00:38 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:25.890 19:00:38 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:04:25.890 [2024-11-26 19:00:38.488040] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:25.890 [2024-11-26 19:00:38.488093] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3605573 ] 00:04:26.461 [2024-11-26 19:00:38.793278] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:26.461 [2024-11-26 19:00:38.824912] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:26.726 19:00:39 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:26.727 19:00:39 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:04:26.727 00:04:26.727 19:00:39 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:04:26.727 INFO: shutting down applications... 00:04:26.727 19:00:39 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 3605573 ]] 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 3605573 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 3605573 00:04:26.727 19:00:39 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:04:27.306 19:00:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:04:27.306 19:00:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:27.306 19:00:39 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 3605573 00:04:27.306 19:00:39 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:04:27.306 19:00:39 json_config_extra_key -- json_config/common.sh@43 -- # break 00:04:27.306 19:00:39 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:04:27.306 19:00:39 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:04:27.306 SPDK target shutdown done 00:04:27.306 19:00:39 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:04:27.306 Success 00:04:27.306 00:04:27.306 real 0m1.550s 00:04:27.306 user 0m1.162s 00:04:27.306 sys 0m0.417s 00:04:27.306 19:00:39 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:27.306 19:00:39 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:04:27.306 ************************************ 00:04:27.306 END TEST json_config_extra_key 00:04:27.306 ************************************ 00:04:27.306 19:00:39 -- spdk/autotest.sh@161 -- # run_test alias_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:04:27.306 19:00:39 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:27.306 19:00:39 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:27.306 19:00:39 -- common/autotest_common.sh@10 -- # set +x 00:04:27.306 ************************************ 00:04:27.306 START TEST alias_rpc 00:04:27.306 ************************************ 00:04:27.306 19:00:39 alias_rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:04:27.568 * Looking for test storage... 00:04:27.568 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc 00:04:27.568 19:00:39 alias_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:27.568 19:00:39 alias_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:04:27.568 19:00:39 alias_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:27.568 19:00:40 alias_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@345 -- # : 1 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:27.568 19:00:40 alias_rpc -- scripts/common.sh@368 -- # return 0 00:04:27.568 19:00:40 alias_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:27.568 19:00:40 alias_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:27.568 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:27.569 --rc genhtml_branch_coverage=1 00:04:27.569 --rc genhtml_function_coverage=1 00:04:27.569 --rc genhtml_legend=1 00:04:27.569 --rc geninfo_all_blocks=1 00:04:27.569 --rc geninfo_unexecuted_blocks=1 00:04:27.569 00:04:27.569 ' 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:27.569 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:27.569 --rc genhtml_branch_coverage=1 00:04:27.569 --rc genhtml_function_coverage=1 00:04:27.569 --rc genhtml_legend=1 00:04:27.569 --rc geninfo_all_blocks=1 00:04:27.569 --rc geninfo_unexecuted_blocks=1 00:04:27.569 00:04:27.569 ' 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:27.569 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:27.569 --rc genhtml_branch_coverage=1 00:04:27.569 --rc genhtml_function_coverage=1 00:04:27.569 --rc genhtml_legend=1 00:04:27.569 --rc geninfo_all_blocks=1 00:04:27.569 --rc geninfo_unexecuted_blocks=1 00:04:27.569 00:04:27.569 ' 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:27.569 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:27.569 --rc genhtml_branch_coverage=1 00:04:27.569 --rc genhtml_function_coverage=1 00:04:27.569 --rc genhtml_legend=1 00:04:27.569 --rc geninfo_all_blocks=1 00:04:27.569 --rc geninfo_unexecuted_blocks=1 00:04:27.569 00:04:27.569 ' 00:04:27.569 19:00:40 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:04:27.569 19:00:40 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=3605968 00:04:27.569 19:00:40 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:27.569 19:00:40 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 3605968 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 3605968 ']' 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:27.569 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:27.569 19:00:40 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:27.569 [2024-11-26 19:00:40.097809] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:27.569 [2024-11-26 19:00:40.097891] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3605968 ] 00:04:27.569 [2024-11-26 19:00:40.180969] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:27.831 [2024-11-26 19:00:40.223800] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:28.405 19:00:40 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:28.405 19:00:40 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:04:28.405 19:00:40 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_config -i 00:04:28.666 19:00:41 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 3605968 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 3605968 ']' 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 3605968 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3605968 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3605968' 00:04:28.666 killing process with pid 3605968 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@973 -- # kill 3605968 00:04:28.666 19:00:41 alias_rpc -- common/autotest_common.sh@978 -- # wait 3605968 00:04:28.928 00:04:28.928 real 0m1.540s 00:04:28.928 user 0m1.699s 00:04:28.928 sys 0m0.424s 00:04:28.928 19:00:41 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:28.928 19:00:41 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:28.928 ************************************ 00:04:28.928 END TEST alias_rpc 00:04:28.928 ************************************ 00:04:28.928 19:00:41 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:04:28.928 19:00:41 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:04:28.928 19:00:41 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:28.928 19:00:41 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:28.928 19:00:41 -- common/autotest_common.sh@10 -- # set +x 00:04:28.928 ************************************ 00:04:28.928 START TEST spdkcli_tcp 00:04:28.928 ************************************ 00:04:28.928 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:04:28.928 * Looking for test storage... 00:04:29.190 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:29.190 19:00:41 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:29.190 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:29.190 --rc genhtml_branch_coverage=1 00:04:29.190 --rc genhtml_function_coverage=1 00:04:29.190 --rc genhtml_legend=1 00:04:29.190 --rc geninfo_all_blocks=1 00:04:29.190 --rc geninfo_unexecuted_blocks=1 00:04:29.190 00:04:29.190 ' 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:29.190 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:29.190 --rc genhtml_branch_coverage=1 00:04:29.190 --rc genhtml_function_coverage=1 00:04:29.190 --rc genhtml_legend=1 00:04:29.190 --rc geninfo_all_blocks=1 00:04:29.190 --rc geninfo_unexecuted_blocks=1 00:04:29.190 00:04:29.190 ' 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:29.190 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:29.190 --rc genhtml_branch_coverage=1 00:04:29.190 --rc genhtml_function_coverage=1 00:04:29.190 --rc genhtml_legend=1 00:04:29.190 --rc geninfo_all_blocks=1 00:04:29.190 --rc geninfo_unexecuted_blocks=1 00:04:29.190 00:04:29.190 ' 00:04:29.190 19:00:41 spdkcli_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:29.190 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:29.190 --rc genhtml_branch_coverage=1 00:04:29.190 --rc genhtml_function_coverage=1 00:04:29.190 --rc genhtml_legend=1 00:04:29.190 --rc geninfo_all_blocks=1 00:04:29.190 --rc geninfo_unexecuted_blocks=1 00:04:29.190 00:04:29.190 ' 00:04:29.190 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:04:29.190 19:00:41 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:04:29.190 19:00:41 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:04:29.190 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:04:29.190 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:04:29.190 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:04:29.190 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:29.191 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=3606361 00:04:29.191 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 3606361 00:04:29.191 19:00:41 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 3606361 ']' 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:29.191 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:29.191 19:00:41 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:29.191 [2024-11-26 19:00:41.717827] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:29.191 [2024-11-26 19:00:41.717897] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3606361 ] 00:04:29.191 [2024-11-26 19:00:41.795109] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:29.525 [2024-11-26 19:00:41.832920] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:29.525 [2024-11-26 19:00:41.833116] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:30.171 19:00:42 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:30.171 19:00:42 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:04:30.171 19:00:42 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=3606658 00:04:30.171 19:00:42 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:04:30.171 19:00:42 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:04:30.171 [ 00:04:30.171 "bdev_malloc_delete", 00:04:30.171 "bdev_malloc_create", 00:04:30.171 "bdev_null_resize", 00:04:30.171 "bdev_null_delete", 00:04:30.171 "bdev_null_create", 00:04:30.171 "bdev_nvme_cuse_unregister", 00:04:30.171 "bdev_nvme_cuse_register", 00:04:30.171 "bdev_opal_new_user", 00:04:30.171 "bdev_opal_set_lock_state", 00:04:30.171 "bdev_opal_delete", 00:04:30.171 "bdev_opal_get_info", 00:04:30.171 "bdev_opal_create", 00:04:30.171 "bdev_nvme_opal_revert", 00:04:30.171 "bdev_nvme_opal_init", 00:04:30.171 "bdev_nvme_send_cmd", 00:04:30.171 "bdev_nvme_set_keys", 00:04:30.171 "bdev_nvme_get_path_iostat", 00:04:30.171 "bdev_nvme_get_mdns_discovery_info", 00:04:30.171 "bdev_nvme_stop_mdns_discovery", 00:04:30.171 "bdev_nvme_start_mdns_discovery", 00:04:30.171 "bdev_nvme_set_multipath_policy", 00:04:30.171 "bdev_nvme_set_preferred_path", 00:04:30.171 "bdev_nvme_get_io_paths", 00:04:30.171 "bdev_nvme_remove_error_injection", 00:04:30.171 "bdev_nvme_add_error_injection", 00:04:30.171 "bdev_nvme_get_discovery_info", 00:04:30.171 "bdev_nvme_stop_discovery", 00:04:30.171 "bdev_nvme_start_discovery", 00:04:30.171 "bdev_nvme_get_controller_health_info", 00:04:30.171 "bdev_nvme_disable_controller", 00:04:30.171 "bdev_nvme_enable_controller", 00:04:30.171 "bdev_nvme_reset_controller", 00:04:30.171 "bdev_nvme_get_transport_statistics", 00:04:30.171 "bdev_nvme_apply_firmware", 00:04:30.171 "bdev_nvme_detach_controller", 00:04:30.171 "bdev_nvme_get_controllers", 00:04:30.171 "bdev_nvme_attach_controller", 00:04:30.171 "bdev_nvme_set_hotplug", 00:04:30.171 "bdev_nvme_set_options", 00:04:30.171 "bdev_passthru_delete", 00:04:30.171 "bdev_passthru_create", 00:04:30.171 "bdev_lvol_set_parent_bdev", 00:04:30.171 "bdev_lvol_set_parent", 00:04:30.171 "bdev_lvol_check_shallow_copy", 00:04:30.171 "bdev_lvol_start_shallow_copy", 00:04:30.171 "bdev_lvol_grow_lvstore", 00:04:30.171 "bdev_lvol_get_lvols", 00:04:30.171 "bdev_lvol_get_lvstores", 00:04:30.171 "bdev_lvol_delete", 00:04:30.171 "bdev_lvol_set_read_only", 00:04:30.171 "bdev_lvol_resize", 00:04:30.171 "bdev_lvol_decouple_parent", 00:04:30.171 "bdev_lvol_inflate", 00:04:30.171 "bdev_lvol_rename", 00:04:30.171 "bdev_lvol_clone_bdev", 00:04:30.171 "bdev_lvol_clone", 00:04:30.171 "bdev_lvol_snapshot", 00:04:30.171 "bdev_lvol_create", 00:04:30.171 "bdev_lvol_delete_lvstore", 00:04:30.172 "bdev_lvol_rename_lvstore", 00:04:30.172 "bdev_lvol_create_lvstore", 00:04:30.172 "bdev_raid_set_options", 00:04:30.172 "bdev_raid_remove_base_bdev", 00:04:30.172 "bdev_raid_add_base_bdev", 00:04:30.172 "bdev_raid_delete", 00:04:30.172 "bdev_raid_create", 00:04:30.172 "bdev_raid_get_bdevs", 00:04:30.172 "bdev_error_inject_error", 00:04:30.172 "bdev_error_delete", 00:04:30.172 "bdev_error_create", 00:04:30.172 "bdev_split_delete", 00:04:30.172 "bdev_split_create", 00:04:30.172 "bdev_delay_delete", 00:04:30.172 "bdev_delay_create", 00:04:30.172 "bdev_delay_update_latency", 00:04:30.172 "bdev_zone_block_delete", 00:04:30.172 "bdev_zone_block_create", 00:04:30.172 "blobfs_create", 00:04:30.172 "blobfs_detect", 00:04:30.172 "blobfs_set_cache_size", 00:04:30.172 "bdev_aio_delete", 00:04:30.172 "bdev_aio_rescan", 00:04:30.172 "bdev_aio_create", 00:04:30.172 "bdev_ftl_set_property", 00:04:30.172 "bdev_ftl_get_properties", 00:04:30.172 "bdev_ftl_get_stats", 00:04:30.172 "bdev_ftl_unmap", 00:04:30.172 "bdev_ftl_unload", 00:04:30.172 "bdev_ftl_delete", 00:04:30.172 "bdev_ftl_load", 00:04:30.172 "bdev_ftl_create", 00:04:30.172 "bdev_virtio_attach_controller", 00:04:30.172 "bdev_virtio_scsi_get_devices", 00:04:30.172 "bdev_virtio_detach_controller", 00:04:30.172 "bdev_virtio_blk_set_hotplug", 00:04:30.172 "bdev_iscsi_delete", 00:04:30.172 "bdev_iscsi_create", 00:04:30.172 "bdev_iscsi_set_options", 00:04:30.172 "accel_error_inject_error", 00:04:30.172 "ioat_scan_accel_module", 00:04:30.172 "dsa_scan_accel_module", 00:04:30.172 "iaa_scan_accel_module", 00:04:30.172 "vfu_virtio_create_fs_endpoint", 00:04:30.172 "vfu_virtio_create_scsi_endpoint", 00:04:30.172 "vfu_virtio_scsi_remove_target", 00:04:30.172 "vfu_virtio_scsi_add_target", 00:04:30.172 "vfu_virtio_create_blk_endpoint", 00:04:30.172 "vfu_virtio_delete_endpoint", 00:04:30.172 "keyring_file_remove_key", 00:04:30.172 "keyring_file_add_key", 00:04:30.172 "keyring_linux_set_options", 00:04:30.172 "fsdev_aio_delete", 00:04:30.172 "fsdev_aio_create", 00:04:30.172 "iscsi_get_histogram", 00:04:30.172 "iscsi_enable_histogram", 00:04:30.172 "iscsi_set_options", 00:04:30.172 "iscsi_get_auth_groups", 00:04:30.172 "iscsi_auth_group_remove_secret", 00:04:30.172 "iscsi_auth_group_add_secret", 00:04:30.172 "iscsi_delete_auth_group", 00:04:30.172 "iscsi_create_auth_group", 00:04:30.172 "iscsi_set_discovery_auth", 00:04:30.172 "iscsi_get_options", 00:04:30.172 "iscsi_target_node_request_logout", 00:04:30.172 "iscsi_target_node_set_redirect", 00:04:30.172 "iscsi_target_node_set_auth", 00:04:30.172 "iscsi_target_node_add_lun", 00:04:30.172 "iscsi_get_stats", 00:04:30.172 "iscsi_get_connections", 00:04:30.172 "iscsi_portal_group_set_auth", 00:04:30.172 "iscsi_start_portal_group", 00:04:30.172 "iscsi_delete_portal_group", 00:04:30.172 "iscsi_create_portal_group", 00:04:30.172 "iscsi_get_portal_groups", 00:04:30.172 "iscsi_delete_target_node", 00:04:30.172 "iscsi_target_node_remove_pg_ig_maps", 00:04:30.172 "iscsi_target_node_add_pg_ig_maps", 00:04:30.172 "iscsi_create_target_node", 00:04:30.172 "iscsi_get_target_nodes", 00:04:30.172 "iscsi_delete_initiator_group", 00:04:30.172 "iscsi_initiator_group_remove_initiators", 00:04:30.172 "iscsi_initiator_group_add_initiators", 00:04:30.172 "iscsi_create_initiator_group", 00:04:30.172 "iscsi_get_initiator_groups", 00:04:30.172 "nvmf_set_crdt", 00:04:30.172 "nvmf_set_config", 00:04:30.172 "nvmf_set_max_subsystems", 00:04:30.172 "nvmf_stop_mdns_prr", 00:04:30.172 "nvmf_publish_mdns_prr", 00:04:30.172 "nvmf_subsystem_get_listeners", 00:04:30.172 "nvmf_subsystem_get_qpairs", 00:04:30.172 "nvmf_subsystem_get_controllers", 00:04:30.172 "nvmf_get_stats", 00:04:30.172 "nvmf_get_transports", 00:04:30.172 "nvmf_create_transport", 00:04:30.172 "nvmf_get_targets", 00:04:30.172 "nvmf_delete_target", 00:04:30.172 "nvmf_create_target", 00:04:30.172 "nvmf_subsystem_allow_any_host", 00:04:30.172 "nvmf_subsystem_set_keys", 00:04:30.172 "nvmf_subsystem_remove_host", 00:04:30.172 "nvmf_subsystem_add_host", 00:04:30.172 "nvmf_ns_remove_host", 00:04:30.172 "nvmf_ns_add_host", 00:04:30.172 "nvmf_subsystem_remove_ns", 00:04:30.172 "nvmf_subsystem_set_ns_ana_group", 00:04:30.172 "nvmf_subsystem_add_ns", 00:04:30.172 "nvmf_subsystem_listener_set_ana_state", 00:04:30.172 "nvmf_discovery_get_referrals", 00:04:30.172 "nvmf_discovery_remove_referral", 00:04:30.172 "nvmf_discovery_add_referral", 00:04:30.172 "nvmf_subsystem_remove_listener", 00:04:30.172 "nvmf_subsystem_add_listener", 00:04:30.172 "nvmf_delete_subsystem", 00:04:30.172 "nvmf_create_subsystem", 00:04:30.172 "nvmf_get_subsystems", 00:04:30.172 "env_dpdk_get_mem_stats", 00:04:30.172 "nbd_get_disks", 00:04:30.172 "nbd_stop_disk", 00:04:30.172 "nbd_start_disk", 00:04:30.172 "ublk_recover_disk", 00:04:30.172 "ublk_get_disks", 00:04:30.172 "ublk_stop_disk", 00:04:30.172 "ublk_start_disk", 00:04:30.172 "ublk_destroy_target", 00:04:30.172 "ublk_create_target", 00:04:30.172 "virtio_blk_create_transport", 00:04:30.172 "virtio_blk_get_transports", 00:04:30.172 "vhost_controller_set_coalescing", 00:04:30.172 "vhost_get_controllers", 00:04:30.172 "vhost_delete_controller", 00:04:30.172 "vhost_create_blk_controller", 00:04:30.172 "vhost_scsi_controller_remove_target", 00:04:30.172 "vhost_scsi_controller_add_target", 00:04:30.172 "vhost_start_scsi_controller", 00:04:30.172 "vhost_create_scsi_controller", 00:04:30.172 "thread_set_cpumask", 00:04:30.172 "scheduler_set_options", 00:04:30.172 "framework_get_governor", 00:04:30.172 "framework_get_scheduler", 00:04:30.172 "framework_set_scheduler", 00:04:30.172 "framework_get_reactors", 00:04:30.172 "thread_get_io_channels", 00:04:30.172 "thread_get_pollers", 00:04:30.172 "thread_get_stats", 00:04:30.172 "framework_monitor_context_switch", 00:04:30.172 "spdk_kill_instance", 00:04:30.172 "log_enable_timestamps", 00:04:30.172 "log_get_flags", 00:04:30.172 "log_clear_flag", 00:04:30.172 "log_set_flag", 00:04:30.172 "log_get_level", 00:04:30.172 "log_set_level", 00:04:30.172 "log_get_print_level", 00:04:30.172 "log_set_print_level", 00:04:30.172 "framework_enable_cpumask_locks", 00:04:30.172 "framework_disable_cpumask_locks", 00:04:30.172 "framework_wait_init", 00:04:30.172 "framework_start_init", 00:04:30.172 "scsi_get_devices", 00:04:30.172 "bdev_get_histogram", 00:04:30.172 "bdev_enable_histogram", 00:04:30.172 "bdev_set_qos_limit", 00:04:30.172 "bdev_set_qd_sampling_period", 00:04:30.172 "bdev_get_bdevs", 00:04:30.172 "bdev_reset_iostat", 00:04:30.172 "bdev_get_iostat", 00:04:30.172 "bdev_examine", 00:04:30.172 "bdev_wait_for_examine", 00:04:30.172 "bdev_set_options", 00:04:30.172 "accel_get_stats", 00:04:30.172 "accel_set_options", 00:04:30.172 "accel_set_driver", 00:04:30.172 "accel_crypto_key_destroy", 00:04:30.172 "accel_crypto_keys_get", 00:04:30.172 "accel_crypto_key_create", 00:04:30.172 "accel_assign_opc", 00:04:30.172 "accel_get_module_info", 00:04:30.172 "accel_get_opc_assignments", 00:04:30.172 "vmd_rescan", 00:04:30.172 "vmd_remove_device", 00:04:30.172 "vmd_enable", 00:04:30.172 "sock_get_default_impl", 00:04:30.172 "sock_set_default_impl", 00:04:30.172 "sock_impl_set_options", 00:04:30.172 "sock_impl_get_options", 00:04:30.172 "iobuf_get_stats", 00:04:30.172 "iobuf_set_options", 00:04:30.172 "keyring_get_keys", 00:04:30.172 "vfu_tgt_set_base_path", 00:04:30.172 "framework_get_pci_devices", 00:04:30.172 "framework_get_config", 00:04:30.172 "framework_get_subsystems", 00:04:30.172 "fsdev_set_opts", 00:04:30.172 "fsdev_get_opts", 00:04:30.172 "trace_get_info", 00:04:30.172 "trace_get_tpoint_group_mask", 00:04:30.172 "trace_disable_tpoint_group", 00:04:30.172 "trace_enable_tpoint_group", 00:04:30.172 "trace_clear_tpoint_mask", 00:04:30.172 "trace_set_tpoint_mask", 00:04:30.172 "notify_get_notifications", 00:04:30.172 "notify_get_types", 00:04:30.172 "spdk_get_version", 00:04:30.172 "rpc_get_methods" 00:04:30.172 ] 00:04:30.172 19:00:42 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:30.172 19:00:42 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:04:30.172 19:00:42 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 3606361 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 3606361 ']' 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 3606361 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3606361 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3606361' 00:04:30.172 killing process with pid 3606361 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 3606361 00:04:30.172 19:00:42 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 3606361 00:04:30.433 00:04:30.433 real 0m1.538s 00:04:30.433 user 0m2.798s 00:04:30.433 sys 0m0.471s 00:04:30.433 19:00:42 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:30.433 19:00:42 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:30.433 ************************************ 00:04:30.433 END TEST spdkcli_tcp 00:04:30.433 ************************************ 00:04:30.433 19:00:43 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:04:30.433 19:00:43 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:30.433 19:00:43 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:30.433 19:00:43 -- common/autotest_common.sh@10 -- # set +x 00:04:30.693 ************************************ 00:04:30.693 START TEST dpdk_mem_utility 00:04:30.693 ************************************ 00:04:30.693 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:04:30.693 * Looking for test storage... 00:04:30.693 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility 00:04:30.693 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:30.693 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lcov --version 00:04:30.693 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:30.693 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:30.693 19:00:43 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:30.694 19:00:43 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:30.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:30.694 --rc genhtml_branch_coverage=1 00:04:30.694 --rc genhtml_function_coverage=1 00:04:30.694 --rc genhtml_legend=1 00:04:30.694 --rc geninfo_all_blocks=1 00:04:30.694 --rc geninfo_unexecuted_blocks=1 00:04:30.694 00:04:30.694 ' 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:30.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:30.694 --rc genhtml_branch_coverage=1 00:04:30.694 --rc genhtml_function_coverage=1 00:04:30.694 --rc genhtml_legend=1 00:04:30.694 --rc geninfo_all_blocks=1 00:04:30.694 --rc geninfo_unexecuted_blocks=1 00:04:30.694 00:04:30.694 ' 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:30.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:30.694 --rc genhtml_branch_coverage=1 00:04:30.694 --rc genhtml_function_coverage=1 00:04:30.694 --rc genhtml_legend=1 00:04:30.694 --rc geninfo_all_blocks=1 00:04:30.694 --rc geninfo_unexecuted_blocks=1 00:04:30.694 00:04:30.694 ' 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:30.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:30.694 --rc genhtml_branch_coverage=1 00:04:30.694 --rc genhtml_function_coverage=1 00:04:30.694 --rc genhtml_legend=1 00:04:30.694 --rc geninfo_all_blocks=1 00:04:30.694 --rc geninfo_unexecuted_blocks=1 00:04:30.694 00:04:30.694 ' 00:04:30.694 19:00:43 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:04:30.694 19:00:43 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:30.694 19:00:43 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=3606787 00:04:30.694 19:00:43 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 3606787 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 3606787 ']' 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:30.694 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:30.694 19:00:43 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:04:30.694 [2024-11-26 19:00:43.311284] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:30.694 [2024-11-26 19:00:43.311340] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3606787 ] 00:04:30.954 [2024-11-26 19:00:43.388179] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:30.954 [2024-11-26 19:00:43.424245] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:31.578 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:31.578 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:04:31.578 19:00:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:04:31.578 19:00:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:04:31.578 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:31.578 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:04:31.578 { 00:04:31.578 "filename": "/tmp/spdk_mem_dump.txt" 00:04:31.578 } 00:04:31.578 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:31.578 19:00:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:04:31.578 DPDK memory size 818.000000 MiB in 1 heap(s) 00:04:31.578 1 heaps totaling size 818.000000 MiB 00:04:31.578 size: 818.000000 MiB heap id: 0 00:04:31.578 end heaps---------- 00:04:31.578 9 mempools totaling size 603.782043 MiB 00:04:31.578 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:04:31.578 size: 158.602051 MiB name: PDU_data_out_Pool 00:04:31.578 size: 100.555481 MiB name: bdev_io_3606787 00:04:31.578 size: 50.003479 MiB name: msgpool_3606787 00:04:31.578 size: 36.509338 MiB name: fsdev_io_3606787 00:04:31.578 size: 21.763794 MiB name: PDU_Pool 00:04:31.578 size: 19.513306 MiB name: SCSI_TASK_Pool 00:04:31.578 size: 4.133484 MiB name: evtpool_3606787 00:04:31.578 size: 0.026123 MiB name: Session_Pool 00:04:31.578 end mempools------- 00:04:31.578 6 memzones totaling size 4.142822 MiB 00:04:31.578 size: 1.000366 MiB name: RG_ring_0_3606787 00:04:31.578 size: 1.000366 MiB name: RG_ring_1_3606787 00:04:31.578 size: 1.000366 MiB name: RG_ring_4_3606787 00:04:31.578 size: 1.000366 MiB name: RG_ring_5_3606787 00:04:31.578 size: 0.125366 MiB name: RG_ring_2_3606787 00:04:31.578 size: 0.015991 MiB name: RG_ring_3_3606787 00:04:31.578 end memzones------- 00:04:31.578 19:00:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py -m 0 00:04:31.839 heap id: 0 total size: 818.000000 MiB number of busy elements: 44 number of free elements: 15 00:04:31.839 list of free elements. size: 10.852478 MiB 00:04:31.839 element at address: 0x200019200000 with size: 0.999878 MiB 00:04:31.839 element at address: 0x200019400000 with size: 0.999878 MiB 00:04:31.839 element at address: 0x200000400000 with size: 0.998535 MiB 00:04:31.839 element at address: 0x200032000000 with size: 0.994446 MiB 00:04:31.839 element at address: 0x200006400000 with size: 0.959839 MiB 00:04:31.839 element at address: 0x200012c00000 with size: 0.944275 MiB 00:04:31.839 element at address: 0x200019600000 with size: 0.936584 MiB 00:04:31.839 element at address: 0x200000200000 with size: 0.717346 MiB 00:04:31.839 element at address: 0x20001ae00000 with size: 0.582886 MiB 00:04:31.839 element at address: 0x200000c00000 with size: 0.495422 MiB 00:04:31.839 element at address: 0x20000a600000 with size: 0.490723 MiB 00:04:31.839 element at address: 0x200019800000 with size: 0.485657 MiB 00:04:31.839 element at address: 0x200003e00000 with size: 0.481934 MiB 00:04:31.839 element at address: 0x200028200000 with size: 0.410034 MiB 00:04:31.839 element at address: 0x200000800000 with size: 0.355042 MiB 00:04:31.839 list of standard malloc elements. size: 199.218628 MiB 00:04:31.839 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:04:31.839 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:04:31.839 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:04:31.839 element at address: 0x2000194fff80 with size: 1.000122 MiB 00:04:31.839 element at address: 0x2000196fff80 with size: 1.000122 MiB 00:04:31.839 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:04:31.839 element at address: 0x2000196eff00 with size: 0.062622 MiB 00:04:31.839 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:04:31.839 element at address: 0x2000196efdc0 with size: 0.000305 MiB 00:04:31.839 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000004ffb80 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000085ae40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000085b040 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000085f300 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000087f680 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200000cff000 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200003efb980 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200012cf1bc0 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000196efc40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000196efd00 with size: 0.000183 MiB 00:04:31.839 element at address: 0x2000198bc740 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20001ae95380 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20001ae95440 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200028268f80 with size: 0.000183 MiB 00:04:31.839 element at address: 0x200028269040 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20002826fc40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20002826fe40 with size: 0.000183 MiB 00:04:31.839 element at address: 0x20002826ff00 with size: 0.000183 MiB 00:04:31.839 list of memzone associated elements. size: 607.928894 MiB 00:04:31.839 element at address: 0x20001ae95500 with size: 211.416748 MiB 00:04:31.839 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:04:31.839 element at address: 0x20002826ffc0 with size: 157.562561 MiB 00:04:31.839 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:04:31.839 element at address: 0x200012df1e80 with size: 100.055054 MiB 00:04:31.839 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_3606787_0 00:04:31.839 element at address: 0x200000dff380 with size: 48.003052 MiB 00:04:31.839 associated memzone info: size: 48.002930 MiB name: MP_msgpool_3606787_0 00:04:31.839 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:04:31.839 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_3606787_0 00:04:31.839 element at address: 0x2000199be940 with size: 20.255554 MiB 00:04:31.839 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:04:31.839 element at address: 0x2000321feb40 with size: 18.005066 MiB 00:04:31.839 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:04:31.839 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:04:31.839 associated memzone info: size: 3.000122 MiB name: MP_evtpool_3606787_0 00:04:31.839 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:04:31.839 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_3606787 00:04:31.839 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:04:31.839 associated memzone info: size: 1.007996 MiB name: MP_evtpool_3606787 00:04:31.839 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:04:31.839 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:04:31.839 element at address: 0x2000198bc800 with size: 1.008118 MiB 00:04:31.839 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:04:31.839 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:04:31.839 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:04:31.839 element at address: 0x200003efba40 with size: 1.008118 MiB 00:04:31.839 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:04:31.839 element at address: 0x200000cff180 with size: 1.000488 MiB 00:04:31.840 associated memzone info: size: 1.000366 MiB name: RG_ring_0_3606787 00:04:31.840 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:04:31.840 associated memzone info: size: 1.000366 MiB name: RG_ring_1_3606787 00:04:31.840 element at address: 0x200012cf1c80 with size: 1.000488 MiB 00:04:31.840 associated memzone info: size: 1.000366 MiB name: RG_ring_4_3606787 00:04:31.840 element at address: 0x2000320fe940 with size: 1.000488 MiB 00:04:31.840 associated memzone info: size: 1.000366 MiB name: RG_ring_5_3606787 00:04:31.840 element at address: 0x20000087f740 with size: 0.500488 MiB 00:04:31.840 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_3606787 00:04:31.840 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:04:31.840 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_3606787 00:04:31.840 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:04:31.840 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:04:31.840 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:04:31.840 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:04:31.840 element at address: 0x20001987c540 with size: 0.250488 MiB 00:04:31.840 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:04:31.840 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:04:31.840 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_3606787 00:04:31.840 element at address: 0x20000085f3c0 with size: 0.125488 MiB 00:04:31.840 associated memzone info: size: 0.125366 MiB name: RG_ring_2_3606787 00:04:31.840 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:04:31.840 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:04:31.840 element at address: 0x200028269100 with size: 0.023743 MiB 00:04:31.840 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:04:31.840 element at address: 0x20000085b100 with size: 0.016113 MiB 00:04:31.840 associated memzone info: size: 0.015991 MiB name: RG_ring_3_3606787 00:04:31.840 element at address: 0x20002826f240 with size: 0.002441 MiB 00:04:31.840 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:04:31.840 element at address: 0x2000004ffc40 with size: 0.000305 MiB 00:04:31.840 associated memzone info: size: 0.000183 MiB name: MP_msgpool_3606787 00:04:31.840 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:04:31.840 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_3606787 00:04:31.840 element at address: 0x20000085af00 with size: 0.000305 MiB 00:04:31.840 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_3606787 00:04:31.840 element at address: 0x20002826fd00 with size: 0.000305 MiB 00:04:31.840 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:04:31.840 19:00:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:04:31.840 19:00:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 3606787 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 3606787 ']' 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 3606787 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3606787 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3606787' 00:04:31.840 killing process with pid 3606787 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 3606787 00:04:31.840 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 3606787 00:04:32.101 00:04:32.101 real 0m1.408s 00:04:32.101 user 0m1.521s 00:04:32.101 sys 0m0.372s 00:04:32.101 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:32.101 19:00:44 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:04:32.101 ************************************ 00:04:32.101 END TEST dpdk_mem_utility 00:04:32.101 ************************************ 00:04:32.101 19:00:44 -- spdk/autotest.sh@168 -- # run_test event /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:04:32.101 19:00:44 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:32.101 19:00:44 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:32.101 19:00:44 -- common/autotest_common.sh@10 -- # set +x 00:04:32.101 ************************************ 00:04:32.101 START TEST event 00:04:32.101 ************************************ 00:04:32.101 19:00:44 event -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:04:32.101 * Looking for test storage... 00:04:32.101 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:04:32.101 19:00:44 event -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:32.101 19:00:44 event -- common/autotest_common.sh@1693 -- # lcov --version 00:04:32.101 19:00:44 event -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:32.361 19:00:44 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:32.361 19:00:44 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:32.361 19:00:44 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:32.361 19:00:44 event -- scripts/common.sh@336 -- # IFS=.-: 00:04:32.361 19:00:44 event -- scripts/common.sh@336 -- # read -ra ver1 00:04:32.361 19:00:44 event -- scripts/common.sh@337 -- # IFS=.-: 00:04:32.361 19:00:44 event -- scripts/common.sh@337 -- # read -ra ver2 00:04:32.361 19:00:44 event -- scripts/common.sh@338 -- # local 'op=<' 00:04:32.361 19:00:44 event -- scripts/common.sh@340 -- # ver1_l=2 00:04:32.361 19:00:44 event -- scripts/common.sh@341 -- # ver2_l=1 00:04:32.361 19:00:44 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:32.361 19:00:44 event -- scripts/common.sh@344 -- # case "$op" in 00:04:32.361 19:00:44 event -- scripts/common.sh@345 -- # : 1 00:04:32.361 19:00:44 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:32.361 19:00:44 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:32.361 19:00:44 event -- scripts/common.sh@365 -- # decimal 1 00:04:32.361 19:00:44 event -- scripts/common.sh@353 -- # local d=1 00:04:32.361 19:00:44 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:32.361 19:00:44 event -- scripts/common.sh@355 -- # echo 1 00:04:32.361 19:00:44 event -- scripts/common.sh@365 -- # ver1[v]=1 00:04:32.361 19:00:44 event -- scripts/common.sh@366 -- # decimal 2 00:04:32.361 19:00:44 event -- scripts/common.sh@353 -- # local d=2 00:04:32.361 19:00:44 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:32.361 19:00:44 event -- scripts/common.sh@355 -- # echo 2 00:04:32.361 19:00:44 event -- scripts/common.sh@366 -- # ver2[v]=2 00:04:32.361 19:00:44 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:32.361 19:00:44 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:32.361 19:00:44 event -- scripts/common.sh@368 -- # return 0 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:32.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:32.361 --rc genhtml_branch_coverage=1 00:04:32.361 --rc genhtml_function_coverage=1 00:04:32.361 --rc genhtml_legend=1 00:04:32.361 --rc geninfo_all_blocks=1 00:04:32.361 --rc geninfo_unexecuted_blocks=1 00:04:32.361 00:04:32.361 ' 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:32.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:32.361 --rc genhtml_branch_coverage=1 00:04:32.361 --rc genhtml_function_coverage=1 00:04:32.361 --rc genhtml_legend=1 00:04:32.361 --rc geninfo_all_blocks=1 00:04:32.361 --rc geninfo_unexecuted_blocks=1 00:04:32.361 00:04:32.361 ' 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:32.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:32.361 --rc genhtml_branch_coverage=1 00:04:32.361 --rc genhtml_function_coverage=1 00:04:32.361 --rc genhtml_legend=1 00:04:32.361 --rc geninfo_all_blocks=1 00:04:32.361 --rc geninfo_unexecuted_blocks=1 00:04:32.361 00:04:32.361 ' 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:32.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:32.361 --rc genhtml_branch_coverage=1 00:04:32.361 --rc genhtml_function_coverage=1 00:04:32.361 --rc genhtml_legend=1 00:04:32.361 --rc geninfo_all_blocks=1 00:04:32.361 --rc geninfo_unexecuted_blocks=1 00:04:32.361 00:04:32.361 ' 00:04:32.361 19:00:44 event -- event/event.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/nbd_common.sh 00:04:32.361 19:00:44 event -- bdev/nbd_common.sh@6 -- # set -e 00:04:32.361 19:00:44 event -- event/event.sh@45 -- # run_test event_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:04:32.361 19:00:44 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:32.361 19:00:44 event -- common/autotest_common.sh@10 -- # set +x 00:04:32.361 ************************************ 00:04:32.361 START TEST event_perf 00:04:32.361 ************************************ 00:04:32.361 19:00:44 event.event_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:04:32.361 Running I/O for 1 seconds...[2024-11-26 19:00:44.816932] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:32.361 [2024-11-26 19:00:44.817009] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3607183 ] 00:04:32.361 [2024-11-26 19:00:44.898438] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:04:32.362 [2024-11-26 19:00:44.938544] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:32.362 [2024-11-26 19:00:44.938660] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:04:32.362 [2024-11-26 19:00:44.938833] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:32.362 Running I/O for 1 seconds...[2024-11-26 19:00:44.938833] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:04:33.746 00:04:33.747 lcore 0: 181978 00:04:33.747 lcore 1: 181977 00:04:33.747 lcore 2: 181974 00:04:33.747 lcore 3: 181977 00:04:33.747 done. 00:04:33.747 00:04:33.747 real 0m1.177s 00:04:33.747 user 0m4.097s 00:04:33.747 sys 0m0.077s 00:04:33.747 19:00:45 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:33.747 19:00:45 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:04:33.747 ************************************ 00:04:33.747 END TEST event_perf 00:04:33.747 ************************************ 00:04:33.747 19:00:46 event -- event/event.sh@46 -- # run_test event_reactor /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:04:33.747 19:00:46 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:04:33.747 19:00:46 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:33.747 19:00:46 event -- common/autotest_common.sh@10 -- # set +x 00:04:33.747 ************************************ 00:04:33.747 START TEST event_reactor 00:04:33.747 ************************************ 00:04:33.747 19:00:46 event.event_reactor -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:04:33.747 [2024-11-26 19:00:46.074293] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:33.747 [2024-11-26 19:00:46.074389] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3607543 ] 00:04:33.747 [2024-11-26 19:00:46.156756] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:33.747 [2024-11-26 19:00:46.192394] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:34.688 test_start 00:04:34.688 oneshot 00:04:34.688 tick 100 00:04:34.688 tick 100 00:04:34.688 tick 250 00:04:34.688 tick 100 00:04:34.688 tick 100 00:04:34.688 tick 250 00:04:34.688 tick 100 00:04:34.688 tick 500 00:04:34.688 tick 100 00:04:34.688 tick 100 00:04:34.688 tick 250 00:04:34.688 tick 100 00:04:34.688 tick 100 00:04:34.688 test_end 00:04:34.688 00:04:34.688 real 0m1.172s 00:04:34.688 user 0m1.105s 00:04:34.688 sys 0m0.064s 00:04:34.688 19:00:47 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:34.688 19:00:47 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:04:34.688 ************************************ 00:04:34.688 END TEST event_reactor 00:04:34.688 ************************************ 00:04:34.688 19:00:47 event -- event/event.sh@47 -- # run_test event_reactor_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:04:34.688 19:00:47 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:04:34.688 19:00:47 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:34.688 19:00:47 event -- common/autotest_common.sh@10 -- # set +x 00:04:34.688 ************************************ 00:04:34.688 START TEST event_reactor_perf 00:04:34.688 ************************************ 00:04:34.688 19:00:47 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:04:34.948 [2024-11-26 19:00:47.323462] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:34.948 [2024-11-26 19:00:47.323559] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3607726 ] 00:04:34.948 [2024-11-26 19:00:47.405400] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:34.948 [2024-11-26 19:00:47.440968] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:35.887 test_start 00:04:35.887 test_end 00:04:35.887 Performance: 370894 events per second 00:04:35.887 00:04:35.887 real 0m1.171s 00:04:35.887 user 0m1.099s 00:04:35.887 sys 0m0.068s 00:04:35.887 19:00:48 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:35.887 19:00:48 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:04:35.887 ************************************ 00:04:35.887 END TEST event_reactor_perf 00:04:35.887 ************************************ 00:04:36.148 19:00:48 event -- event/event.sh@49 -- # uname -s 00:04:36.148 19:00:48 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:04:36.148 19:00:48 event -- event/event.sh@50 -- # run_test event_scheduler /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:04:36.148 19:00:48 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:36.148 19:00:48 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:36.148 19:00:48 event -- common/autotest_common.sh@10 -- # set +x 00:04:36.148 ************************************ 00:04:36.148 START TEST event_scheduler 00:04:36.148 ************************************ 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:04:36.148 * Looking for test storage... 00:04:36.148 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1693 -- # lcov --version 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:36.148 19:00:48 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:36.148 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.148 --rc genhtml_branch_coverage=1 00:04:36.148 --rc genhtml_function_coverage=1 00:04:36.148 --rc genhtml_legend=1 00:04:36.148 --rc geninfo_all_blocks=1 00:04:36.148 --rc geninfo_unexecuted_blocks=1 00:04:36.148 00:04:36.148 ' 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:36.148 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.148 --rc genhtml_branch_coverage=1 00:04:36.148 --rc genhtml_function_coverage=1 00:04:36.148 --rc genhtml_legend=1 00:04:36.148 --rc geninfo_all_blocks=1 00:04:36.148 --rc geninfo_unexecuted_blocks=1 00:04:36.148 00:04:36.148 ' 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:36.148 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.148 --rc genhtml_branch_coverage=1 00:04:36.148 --rc genhtml_function_coverage=1 00:04:36.148 --rc genhtml_legend=1 00:04:36.148 --rc geninfo_all_blocks=1 00:04:36.148 --rc geninfo_unexecuted_blocks=1 00:04:36.148 00:04:36.148 ' 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:36.148 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.148 --rc genhtml_branch_coverage=1 00:04:36.148 --rc genhtml_function_coverage=1 00:04:36.148 --rc genhtml_legend=1 00:04:36.148 --rc geninfo_all_blocks=1 00:04:36.148 --rc geninfo_unexecuted_blocks=1 00:04:36.148 00:04:36.148 ' 00:04:36.148 19:00:48 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:04:36.148 19:00:48 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=3607984 00:04:36.148 19:00:48 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:04:36.148 19:00:48 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 3607984 00:04:36.148 19:00:48 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 3607984 ']' 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:36.148 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:36.148 19:00:48 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:36.407 [2024-11-26 19:00:48.804697] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:36.407 [2024-11-26 19:00:48.804755] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3607984 ] 00:04:36.407 [2024-11-26 19:00:48.870236] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:04:36.407 [2024-11-26 19:00:48.903396] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:36.407 [2024-11-26 19:00:48.903552] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:36.407 [2024-11-26 19:00:48.903704] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:04:36.407 [2024-11-26 19:00:48.903706] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:04:36.976 19:00:49 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:36.976 19:00:49 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:04:36.976 19:00:49 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:04:36.976 19:00:49 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:36.976 19:00:49 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 [2024-11-26 19:00:49.601851] dpdk_governor.c: 178:_init: *ERROR*: App core mask contains some but not all of a set of SMT siblings 00:04:37.237 [2024-11-26 19:00:49.601870] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:04:37.237 [2024-11-26 19:00:49.601878] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:04:37.237 [2024-11-26 19:00:49.601882] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:04:37.237 [2024-11-26 19:00:49.601886] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:04:37.237 19:00:49 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:04:37.237 19:00:49 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 [2024-11-26 19:00:49.662837] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:04:37.237 19:00:49 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:04:37.237 19:00:49 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:37.237 19:00:49 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 ************************************ 00:04:37.237 START TEST scheduler_create_thread 00:04:37.237 ************************************ 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 2 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 3 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 4 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 5 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 6 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 7 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 8 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.237 9 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.237 19:00:49 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:37.809 10 00:04:37.809 19:00:50 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:37.809 19:00:50 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:04:37.809 19:00:50 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:37.809 19:00:50 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:39.192 19:00:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:39.192 19:00:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:04:39.192 19:00:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:04:39.192 19:00:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:39.192 19:00:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:40.137 19:00:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:40.137 19:00:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:04:40.137 19:00:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:40.137 19:00:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:40.710 19:00:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:40.710 19:00:53 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:04:40.710 19:00:53 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:04:40.710 19:00:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:40.710 19:00:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:41.652 19:00:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:41.652 00:04:41.652 real 0m4.225s 00:04:41.652 user 0m0.025s 00:04:41.652 sys 0m0.006s 00:04:41.652 19:00:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:41.652 19:00:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:41.652 ************************************ 00:04:41.652 END TEST scheduler_create_thread 00:04:41.652 ************************************ 00:04:41.652 19:00:53 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:04:41.652 19:00:53 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 3607984 00:04:41.652 19:00:53 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 3607984 ']' 00:04:41.652 19:00:53 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 3607984 00:04:41.652 19:00:53 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:04:41.652 19:00:53 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:41.652 19:00:53 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3607984 00:04:41.652 19:00:54 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:04:41.652 19:00:54 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:04:41.652 19:00:54 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3607984' 00:04:41.652 killing process with pid 3607984 00:04:41.652 19:00:54 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 3607984 00:04:41.652 19:00:54 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 3607984 00:04:41.652 [2024-11-26 19:00:54.204077] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:04:41.912 00:04:41.912 real 0m5.806s 00:04:41.912 user 0m12.959s 00:04:41.912 sys 0m0.388s 00:04:41.912 19:00:54 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:41.912 19:00:54 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:41.912 ************************************ 00:04:41.912 END TEST event_scheduler 00:04:41.912 ************************************ 00:04:41.912 19:00:54 event -- event/event.sh@51 -- # modprobe -n nbd 00:04:41.912 19:00:54 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:04:41.912 19:00:54 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:41.912 19:00:54 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:41.912 19:00:54 event -- common/autotest_common.sh@10 -- # set +x 00:04:41.912 ************************************ 00:04:41.912 START TEST app_repeat 00:04:41.912 ************************************ 00:04:41.912 19:00:54 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@19 -- # repeat_pid=3609350 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 3609350' 00:04:41.912 Process app_repeat pid: 3609350 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:04:41.912 spdk_app_start Round 0 00:04:41.912 19:00:54 event.app_repeat -- event/event.sh@25 -- # waitforlisten 3609350 /var/tmp/spdk-nbd.sock 00:04:41.912 19:00:54 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 3609350 ']' 00:04:41.912 19:00:54 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:04:41.912 19:00:54 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:41.912 19:00:54 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:04:41.912 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:04:41.912 19:00:54 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:41.912 19:00:54 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:41.913 [2024-11-26 19:00:54.485445] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:41.913 [2024-11-26 19:00:54.485532] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3609350 ] 00:04:42.172 [2024-11-26 19:00:54.565243] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:42.172 [2024-11-26 19:00:54.601957] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:42.172 [2024-11-26 19:00:54.602115] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:42.172 19:00:54 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:42.172 19:00:54 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:04:42.172 19:00:54 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:42.432 Malloc0 00:04:42.432 19:00:54 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:42.432 Malloc1 00:04:42.432 19:00:55 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:04:42.432 19:00:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:42.433 19:00:55 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:04:42.693 /dev/nbd0 00:04:42.693 19:00:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:04:42.693 19:00:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:42.693 1+0 records in 00:04:42.693 1+0 records out 00:04:42.693 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000144475 s, 28.4 MB/s 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:04:42.693 19:00:55 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:04:42.693 19:00:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:42.693 19:00:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:42.693 19:00:55 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:04:42.954 /dev/nbd1 00:04:42.954 19:00:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:04:42.954 19:00:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:42.954 1+0 records in 00:04:42.954 1+0 records out 00:04:42.954 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000303634 s, 13.5 MB/s 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:04:42.954 19:00:55 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:04:42.954 19:00:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:42.954 19:00:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:42.954 19:00:55 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:42.954 19:00:55 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:42.954 19:00:55 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:43.215 19:00:55 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:04:43.215 { 00:04:43.215 "nbd_device": "/dev/nbd0", 00:04:43.215 "bdev_name": "Malloc0" 00:04:43.216 }, 00:04:43.216 { 00:04:43.216 "nbd_device": "/dev/nbd1", 00:04:43.216 "bdev_name": "Malloc1" 00:04:43.216 } 00:04:43.216 ]' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:04:43.216 { 00:04:43.216 "nbd_device": "/dev/nbd0", 00:04:43.216 "bdev_name": "Malloc0" 00:04:43.216 }, 00:04:43.216 { 00:04:43.216 "nbd_device": "/dev/nbd1", 00:04:43.216 "bdev_name": "Malloc1" 00:04:43.216 } 00:04:43.216 ]' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:04:43.216 /dev/nbd1' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:04:43.216 /dev/nbd1' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:04:43.216 256+0 records in 00:04:43.216 256+0 records out 00:04:43.216 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0124991 s, 83.9 MB/s 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:04:43.216 256+0 records in 00:04:43.216 256+0 records out 00:04:43.216 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0172742 s, 60.7 MB/s 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:04:43.216 256+0 records in 00:04:43.216 256+0 records out 00:04:43.216 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0178247 s, 58.8 MB/s 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:43.216 19:00:55 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:43.478 19:00:55 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:04:43.740 19:00:56 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:44.000 19:00:56 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:04:44.000 19:00:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:04:44.000 19:00:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:44.000 19:00:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:04:44.000 19:00:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:04:44.000 19:00:56 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:04:44.000 19:00:56 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:04:44.001 19:00:56 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:04:44.001 19:00:56 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:04:44.001 19:00:56 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:04:44.001 19:00:56 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:04:44.261 [2024-11-26 19:00:56.680133] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:44.261 [2024-11-26 19:00:56.716171] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:44.261 [2024-11-26 19:00:56.716173] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:44.261 [2024-11-26 19:00:56.747939] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:04:44.261 [2024-11-26 19:00:56.747975] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:04:47.561 19:00:59 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:04:47.561 19:00:59 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:04:47.561 spdk_app_start Round 1 00:04:47.561 19:00:59 event.app_repeat -- event/event.sh@25 -- # waitforlisten 3609350 /var/tmp/spdk-nbd.sock 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 3609350 ']' 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:04:47.561 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:47.561 19:00:59 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:04:47.561 19:00:59 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:47.561 Malloc0 00:04:47.561 19:00:59 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:47.561 Malloc1 00:04:47.561 19:01:00 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:47.561 19:01:00 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:04:47.823 /dev/nbd0 00:04:47.823 19:01:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:04:47.823 19:01:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:47.823 1+0 records in 00:04:47.823 1+0 records out 00:04:47.823 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000230703 s, 17.8 MB/s 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:04:47.823 19:01:00 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:04:47.823 19:01:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:47.823 19:01:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:47.823 19:01:00 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:04:48.086 /dev/nbd1 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:48.086 1+0 records in 00:04:48.086 1+0 records out 00:04:48.086 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000261842 s, 15.6 MB/s 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:04:48.086 19:01:00 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:04:48.086 { 00:04:48.086 "nbd_device": "/dev/nbd0", 00:04:48.086 "bdev_name": "Malloc0" 00:04:48.086 }, 00:04:48.086 { 00:04:48.086 "nbd_device": "/dev/nbd1", 00:04:48.086 "bdev_name": "Malloc1" 00:04:48.086 } 00:04:48.086 ]' 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:04:48.086 { 00:04:48.086 "nbd_device": "/dev/nbd0", 00:04:48.086 "bdev_name": "Malloc0" 00:04:48.086 }, 00:04:48.086 { 00:04:48.086 "nbd_device": "/dev/nbd1", 00:04:48.086 "bdev_name": "Malloc1" 00:04:48.086 } 00:04:48.086 ]' 00:04:48.086 19:01:00 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:04:48.347 /dev/nbd1' 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:04:48.347 /dev/nbd1' 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:04:48.347 256+0 records in 00:04:48.347 256+0 records out 00:04:48.347 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0121244 s, 86.5 MB/s 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:04:48.347 256+0 records in 00:04:48.347 256+0 records out 00:04:48.347 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0176987 s, 59.2 MB/s 00:04:48.347 19:01:00 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:04:48.348 256+0 records in 00:04:48.348 256+0 records out 00:04:48.348 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0176507 s, 59.4 MB/s 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:48.348 19:01:00 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:48.609 19:01:01 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:04:48.869 19:01:01 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:04:48.869 19:01:01 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:04:49.129 19:01:01 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:04:49.129 [2024-11-26 19:01:01.736783] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:49.390 [2024-11-26 19:01:01.773288] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:49.390 [2024-11-26 19:01:01.773292] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:49.390 [2024-11-26 19:01:01.805747] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:04:49.390 [2024-11-26 19:01:01.805781] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:04:52.689 19:01:04 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:04:52.689 19:01:04 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:04:52.689 spdk_app_start Round 2 00:04:52.689 19:01:04 event.app_repeat -- event/event.sh@25 -- # waitforlisten 3609350 /var/tmp/spdk-nbd.sock 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 3609350 ']' 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:04:52.689 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:52.689 19:01:04 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:04:52.689 19:01:04 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:52.689 Malloc0 00:04:52.689 19:01:04 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:52.689 Malloc1 00:04:52.689 19:01:05 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:52.689 19:01:05 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:04:52.951 /dev/nbd0 00:04:52.951 19:01:05 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:04:52.951 19:01:05 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:52.951 1+0 records in 00:04:52.951 1+0 records out 00:04:52.951 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000272787 s, 15.0 MB/s 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:04:52.951 19:01:05 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:04:52.951 19:01:05 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:52.951 19:01:05 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:52.951 19:01:05 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:04:52.951 /dev/nbd1 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:53.214 1+0 records in 00:04:53.214 1+0 records out 00:04:53.214 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000289643 s, 14.1 MB/s 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:04:53.214 19:01:05 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:04:53.214 { 00:04:53.214 "nbd_device": "/dev/nbd0", 00:04:53.214 "bdev_name": "Malloc0" 00:04:53.214 }, 00:04:53.214 { 00:04:53.214 "nbd_device": "/dev/nbd1", 00:04:53.214 "bdev_name": "Malloc1" 00:04:53.214 } 00:04:53.214 ]' 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:04:53.214 { 00:04:53.214 "nbd_device": "/dev/nbd0", 00:04:53.214 "bdev_name": "Malloc0" 00:04:53.214 }, 00:04:53.214 { 00:04:53.214 "nbd_device": "/dev/nbd1", 00:04:53.214 "bdev_name": "Malloc1" 00:04:53.214 } 00:04:53.214 ]' 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:04:53.214 /dev/nbd1' 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:04:53.214 /dev/nbd1' 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:04:53.214 19:01:05 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:04:53.475 19:01:05 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:04:53.475 256+0 records in 00:04:53.475 256+0 records out 00:04:53.476 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.011878 s, 88.3 MB/s 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:04:53.476 256+0 records in 00:04:53.476 256+0 records out 00:04:53.476 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.016578 s, 63.3 MB/s 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:04:53.476 256+0 records in 00:04:53.476 256+0 records out 00:04:53.476 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0309743 s, 33.9 MB/s 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:53.476 19:01:05 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:04:53.737 19:01:06 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:04:53.737 19:01:06 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:04:53.737 19:01:06 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:04:53.737 19:01:06 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:53.737 19:01:06 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:53.737 19:01:06 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:53.738 19:01:06 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:04:53.999 19:01:06 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:04:53.999 19:01:06 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:04:54.259 19:01:06 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:04:54.259 [2024-11-26 19:01:06.830134] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:54.259 [2024-11-26 19:01:06.866494] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:54.259 [2024-11-26 19:01:06.866497] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:54.520 [2024-11-26 19:01:06.898262] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:04:54.520 [2024-11-26 19:01:06.898302] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:04:57.826 19:01:09 event.app_repeat -- event/event.sh@38 -- # waitforlisten 3609350 /var/tmp/spdk-nbd.sock 00:04:57.826 19:01:09 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 3609350 ']' 00:04:57.826 19:01:09 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:04:57.826 19:01:09 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:57.826 19:01:09 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:04:57.826 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:04:57.827 19:01:09 event.app_repeat -- event/event.sh@39 -- # killprocess 3609350 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 3609350 ']' 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 3609350 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3609350 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3609350' 00:04:57.827 killing process with pid 3609350 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@973 -- # kill 3609350 00:04:57.827 19:01:09 event.app_repeat -- common/autotest_common.sh@978 -- # wait 3609350 00:04:57.827 spdk_app_start is called in Round 0. 00:04:57.827 Shutdown signal received, stop current app iteration 00:04:57.827 Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 reinitialization... 00:04:57.827 spdk_app_start is called in Round 1. 00:04:57.827 Shutdown signal received, stop current app iteration 00:04:57.827 Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 reinitialization... 00:04:57.827 spdk_app_start is called in Round 2. 00:04:57.827 Shutdown signal received, stop current app iteration 00:04:57.827 Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 reinitialization... 00:04:57.827 spdk_app_start is called in Round 3. 00:04:57.827 Shutdown signal received, stop current app iteration 00:04:57.827 19:01:10 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:04:57.827 19:01:10 event.app_repeat -- event/event.sh@42 -- # return 0 00:04:57.827 00:04:57.827 real 0m15.602s 00:04:57.827 user 0m33.956s 00:04:57.827 sys 0m2.280s 00:04:57.827 19:01:10 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:57.827 19:01:10 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:57.827 ************************************ 00:04:57.827 END TEST app_repeat 00:04:57.827 ************************************ 00:04:57.827 19:01:10 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:04:57.827 19:01:10 event -- event/event.sh@55 -- # run_test cpu_locks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:04:57.827 19:01:10 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:57.827 19:01:10 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:57.827 19:01:10 event -- common/autotest_common.sh@10 -- # set +x 00:04:57.827 ************************************ 00:04:57.827 START TEST cpu_locks 00:04:57.827 ************************************ 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:04:57.827 * Looking for test storage... 00:04:57.827 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1693 -- # lcov --version 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:57.827 19:01:10 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:57.827 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:57.827 --rc genhtml_branch_coverage=1 00:04:57.827 --rc genhtml_function_coverage=1 00:04:57.827 --rc genhtml_legend=1 00:04:57.827 --rc geninfo_all_blocks=1 00:04:57.827 --rc geninfo_unexecuted_blocks=1 00:04:57.827 00:04:57.827 ' 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:57.827 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:57.827 --rc genhtml_branch_coverage=1 00:04:57.827 --rc genhtml_function_coverage=1 00:04:57.827 --rc genhtml_legend=1 00:04:57.827 --rc geninfo_all_blocks=1 00:04:57.827 --rc geninfo_unexecuted_blocks=1 00:04:57.827 00:04:57.827 ' 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:57.827 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:57.827 --rc genhtml_branch_coverage=1 00:04:57.827 --rc genhtml_function_coverage=1 00:04:57.827 --rc genhtml_legend=1 00:04:57.827 --rc geninfo_all_blocks=1 00:04:57.827 --rc geninfo_unexecuted_blocks=1 00:04:57.827 00:04:57.827 ' 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:57.827 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:57.827 --rc genhtml_branch_coverage=1 00:04:57.827 --rc genhtml_function_coverage=1 00:04:57.827 --rc genhtml_legend=1 00:04:57.827 --rc geninfo_all_blocks=1 00:04:57.827 --rc geninfo_unexecuted_blocks=1 00:04:57.827 00:04:57.827 ' 00:04:57.827 19:01:10 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:04:57.827 19:01:10 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:04:57.827 19:01:10 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:04:57.827 19:01:10 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:57.827 19:01:10 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:57.827 ************************************ 00:04:57.827 START TEST default_locks 00:04:57.827 ************************************ 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=3612616 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 3612616 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 3612616 ']' 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:57.828 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:57.828 19:01:10 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:04:57.828 [2024-11-26 19:01:10.418936] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:57.828 [2024-11-26 19:01:10.418985] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3612616 ] 00:04:58.089 [2024-11-26 19:01:10.497981] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:58.089 [2024-11-26 19:01:10.534721] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:58.663 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:58.663 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:04:58.663 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 3612616 00:04:58.663 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 3612616 00:04:58.663 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:04:58.925 lslocks: write error 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 3612616 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 3612616 ']' 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 3612616 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3612616 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3612616' 00:04:58.925 killing process with pid 3612616 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 3612616 00:04:58.925 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 3612616 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 3612616 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 3612616 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 3612616 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 3612616 ']' 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:59.187 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:04:59.187 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 850: kill: (3612616) - No such process 00:04:59.187 ERROR: process (pid: 3612616) is no longer running 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:04:59.187 00:04:59.187 real 0m1.225s 00:04:59.187 user 0m1.330s 00:04:59.187 sys 0m0.376s 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:59.187 19:01:11 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:04:59.187 ************************************ 00:04:59.187 END TEST default_locks 00:04:59.187 ************************************ 00:04:59.187 19:01:11 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:04:59.187 19:01:11 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:59.187 19:01:11 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:59.187 19:01:11 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:59.187 ************************************ 00:04:59.187 START TEST default_locks_via_rpc 00:04:59.187 ************************************ 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=3612977 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 3612977 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 3612977 ']' 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:59.187 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:59.187 19:01:11 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:59.187 [2024-11-26 19:01:11.717349] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:04:59.187 [2024-11-26 19:01:11.717396] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3612977 ] 00:04:59.187 [2024-11-26 19:01:11.795076] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:59.449 [2024-11-26 19:01:11.830180] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 3612977 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 3612977 00:05:00.024 19:01:12 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 3612977 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 3612977 ']' 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 3612977 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3612977 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3612977' 00:05:00.598 killing process with pid 3612977 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 3612977 00:05:00.598 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 3612977 00:05:00.860 00:05:00.860 real 0m1.650s 00:05:00.860 user 0m1.787s 00:05:00.860 sys 0m0.540s 00:05:00.860 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:00.860 19:01:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:00.860 ************************************ 00:05:00.860 END TEST default_locks_via_rpc 00:05:00.860 ************************************ 00:05:00.860 19:01:13 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:05:00.860 19:01:13 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:00.860 19:01:13 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:00.860 19:01:13 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:00.860 ************************************ 00:05:00.860 START TEST non_locking_app_on_locked_coremask 00:05:00.860 ************************************ 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=3613344 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 3613344 /var/tmp/spdk.sock 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 3613344 ']' 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:00.860 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:00.860 19:01:13 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:00.860 [2024-11-26 19:01:13.421887] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:00.860 [2024-11-26 19:01:13.421937] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3613344 ] 00:05:01.121 [2024-11-26 19:01:13.498012] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:01.121 [2024-11-26 19:01:13.534097] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=3613461 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 3613461 /var/tmp/spdk2.sock 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 3613461 ']' 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:01.695 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:01.695 19:01:14 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:01.695 [2024-11-26 19:01:14.242237] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:01.695 [2024-11-26 19:01:14.242290] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3613461 ] 00:05:01.956 [2024-11-26 19:01:14.363266] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:01.956 [2024-11-26 19:01:14.363299] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:01.956 [2024-11-26 19:01:14.435998] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:02.529 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:02.529 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:02.529 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 3613344 00:05:02.529 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 3613344 00:05:02.529 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:02.790 lslocks: write error 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 3613344 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 3613344 ']' 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 3613344 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3613344 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3613344' 00:05:02.790 killing process with pid 3613344 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 3613344 00:05:02.790 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 3613344 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 3613461 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 3613461 ']' 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 3613461 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3613461 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3613461' 00:05:03.361 killing process with pid 3613461 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 3613461 00:05:03.361 19:01:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 3613461 00:05:03.622 00:05:03.622 real 0m2.668s 00:05:03.622 user 0m2.948s 00:05:03.622 sys 0m0.773s 00:05:03.622 19:01:16 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:03.622 19:01:16 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:03.622 ************************************ 00:05:03.622 END TEST non_locking_app_on_locked_coremask 00:05:03.622 ************************************ 00:05:03.622 19:01:16 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:05:03.622 19:01:16 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:03.622 19:01:16 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:03.622 19:01:16 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:03.622 ************************************ 00:05:03.622 START TEST locking_app_on_unlocked_coremask 00:05:03.622 ************************************ 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=3613939 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 3613939 /var/tmp/spdk.sock 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 3613939 ']' 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:03.622 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:03.622 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:03.622 [2024-11-26 19:01:16.182157] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:03.622 [2024-11-26 19:01:16.182206] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3613939 ] 00:05:03.883 [2024-11-26 19:01:16.260593] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:03.883 [2024-11-26 19:01:16.260621] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:03.883 [2024-11-26 19:01:16.297169] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=3614058 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 3614058 /var/tmp/spdk2.sock 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 3614058 ']' 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:03.883 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:03.883 19:01:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:04.144 [2024-11-26 19:01:16.542855] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:04.144 [2024-11-26 19:01:16.542913] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3614058 ] 00:05:04.144 [2024-11-26 19:01:16.666310] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:04.144 [2024-11-26 19:01:16.738907] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:04.719 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:04.719 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:04.719 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 3614058 00:05:04.719 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 3614058 00:05:04.719 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:04.979 lslocks: write error 00:05:04.979 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 3613939 00:05:04.979 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 3613939 ']' 00:05:04.979 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 3613939 00:05:04.979 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:04.979 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:04.979 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3613939 00:05:05.240 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:05.240 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:05.240 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3613939' 00:05:05.240 killing process with pid 3613939 00:05:05.240 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 3613939 00:05:05.240 19:01:17 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 3613939 00:05:05.499 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 3614058 00:05:05.499 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 3614058 ']' 00:05:05.499 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 3614058 00:05:05.499 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:05.499 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:05.499 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3614058 00:05:05.758 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:05.758 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:05.758 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3614058' 00:05:05.758 killing process with pid 3614058 00:05:05.758 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 3614058 00:05:05.758 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 3614058 00:05:05.758 00:05:05.758 real 0m2.209s 00:05:05.758 user 0m2.441s 00:05:05.758 sys 0m0.778s 00:05:05.758 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:05.758 19:01:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:05.758 ************************************ 00:05:05.758 END TEST locking_app_on_unlocked_coremask 00:05:05.758 ************************************ 00:05:05.758 19:01:18 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:05:05.758 19:01:18 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:05.758 19:01:18 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:05.758 19:01:18 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:06.018 ************************************ 00:05:06.018 START TEST locking_app_on_locked_coremask 00:05:06.018 ************************************ 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=3614429 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 3614429 /var/tmp/spdk.sock 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 3614429 ']' 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:06.018 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:06.018 19:01:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:06.018 [2024-11-26 19:01:18.466224] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:06.018 [2024-11-26 19:01:18.466274] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3614429 ] 00:05:06.018 [2024-11-26 19:01:18.545260] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:06.018 [2024-11-26 19:01:18.583674] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=3614524 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 3614524 /var/tmp/spdk2.sock 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 3614524 /var/tmp/spdk2.sock 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 3614524 /var/tmp/spdk2.sock 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 3614524 ']' 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:06.961 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:06.961 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:06.961 [2024-11-26 19:01:19.310494] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:06.961 [2024-11-26 19:01:19.310548] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3614524 ] 00:05:06.961 [2024-11-26 19:01:19.433313] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 3614429 has claimed it. 00:05:06.961 [2024-11-26 19:01:19.433354] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:05:07.534 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 850: kill: (3614524) - No such process 00:05:07.534 ERROR: process (pid: 3614524) is no longer running 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 3614429 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 3614429 00:05:07.535 19:01:19 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:07.796 lslocks: write error 00:05:07.796 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 3614429 00:05:07.796 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 3614429 ']' 00:05:07.796 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 3614429 00:05:07.796 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:07.796 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:07.796 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3614429 00:05:08.057 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:08.057 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:08.057 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3614429' 00:05:08.057 killing process with pid 3614429 00:05:08.057 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 3614429 00:05:08.057 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 3614429 00:05:08.057 00:05:08.057 real 0m2.225s 00:05:08.057 user 0m2.513s 00:05:08.057 sys 0m0.620s 00:05:08.058 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:08.058 19:01:20 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:08.058 ************************************ 00:05:08.058 END TEST locking_app_on_locked_coremask 00:05:08.058 ************************************ 00:05:08.058 19:01:20 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:05:08.058 19:01:20 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:08.058 19:01:20 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:08.058 19:01:20 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:08.319 ************************************ 00:05:08.319 START TEST locking_overlapped_coremask 00:05:08.319 ************************************ 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=3614811 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 3614811 /var/tmp/spdk.sock 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 3614811 ']' 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:08.319 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:08.319 19:01:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:08.319 [2024-11-26 19:01:20.755337] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:08.319 [2024-11-26 19:01:20.755393] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3614811 ] 00:05:08.319 [2024-11-26 19:01:20.835272] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:08.319 [2024-11-26 19:01:20.876706] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:08.319 [2024-11-26 19:01:20.876842] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:08.319 [2024-11-26 19:01:20.876845] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=3615142 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 3615142 /var/tmp/spdk2.sock 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 3615142 /var/tmp/spdk2.sock 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:09.262 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 3615142 /var/tmp/spdk2.sock 00:05:09.263 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 3615142 ']' 00:05:09.263 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:09.263 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:09.263 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:09.263 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:09.263 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:09.263 19:01:21 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:09.263 [2024-11-26 19:01:21.613287] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:09.263 [2024-11-26 19:01:21.613342] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3615142 ] 00:05:09.263 [2024-11-26 19:01:21.711681] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 3614811 has claimed it. 00:05:09.263 [2024-11-26 19:01:21.711712] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:05:09.835 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 850: kill: (3615142) - No such process 00:05:09.835 ERROR: process (pid: 3615142) is no longer running 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 3614811 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 3614811 ']' 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 3614811 00:05:09.835 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:05:09.836 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:09.836 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3614811 00:05:09.836 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:09.836 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:09.836 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3614811' 00:05:09.836 killing process with pid 3614811 00:05:09.836 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 3614811 00:05:09.836 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 3614811 00:05:10.097 00:05:10.097 real 0m1.801s 00:05:10.097 user 0m5.189s 00:05:10.097 sys 0m0.397s 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:10.097 ************************************ 00:05:10.097 END TEST locking_overlapped_coremask 00:05:10.097 ************************************ 00:05:10.097 19:01:22 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:05:10.097 19:01:22 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:10.097 19:01:22 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:10.097 19:01:22 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:10.097 ************************************ 00:05:10.097 START TEST locking_overlapped_coremask_via_rpc 00:05:10.097 ************************************ 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=3615232 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 3615232 /var/tmp/spdk.sock 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 3615232 ']' 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:10.097 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:10.097 19:01:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:10.097 [2024-11-26 19:01:22.646403] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:10.097 [2024-11-26 19:01:22.646462] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3615232 ] 00:05:10.358 [2024-11-26 19:01:22.730874] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:10.358 [2024-11-26 19:01:22.730907] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:10.358 [2024-11-26 19:01:22.772598] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:10.358 [2024-11-26 19:01:22.772683] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:10.358 [2024-11-26 19:01:22.772685] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=3615518 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 3615518 /var/tmp/spdk2.sock 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 3615518 ']' 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:10.929 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:10.929 19:01:23 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:10.929 [2024-11-26 19:01:23.487555] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:10.929 [2024-11-26 19:01:23.487607] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3615518 ] 00:05:11.188 [2024-11-26 19:01:23.586493] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:11.188 [2024-11-26 19:01:23.586520] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:11.188 [2024-11-26 19:01:23.649903] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:05:11.188 [2024-11-26 19:01:23.649984] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:11.188 [2024-11-26 19:01:23.649986] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:11.760 [2024-11-26 19:01:24.306930] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 3615232 has claimed it. 00:05:11.760 request: 00:05:11.760 { 00:05:11.760 "method": "framework_enable_cpumask_locks", 00:05:11.760 "req_id": 1 00:05:11.760 } 00:05:11.760 Got JSON-RPC error response 00:05:11.760 response: 00:05:11.760 { 00:05:11.760 "code": -32603, 00:05:11.760 "message": "Failed to claim CPU core: 2" 00:05:11.760 } 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 3615232 /var/tmp/spdk.sock 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 3615232 ']' 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:11.760 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:11.760 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 3615518 /var/tmp/spdk2.sock 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 3615518 ']' 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:12.022 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:12.022 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:05:12.307 00:05:12.307 real 0m2.105s 00:05:12.307 user 0m0.875s 00:05:12.307 sys 0m0.145s 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:12.307 19:01:24 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:12.307 ************************************ 00:05:12.307 END TEST locking_overlapped_coremask_via_rpc 00:05:12.307 ************************************ 00:05:12.307 19:01:24 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:05:12.307 19:01:24 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 3615232 ]] 00:05:12.307 19:01:24 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 3615232 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 3615232 ']' 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 3615232 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3615232 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3615232' 00:05:12.307 killing process with pid 3615232 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 3615232 00:05:12.307 19:01:24 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 3615232 00:05:12.581 19:01:25 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 3615518 ]] 00:05:12.581 19:01:25 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 3615518 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 3615518 ']' 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 3615518 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3615518 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3615518' 00:05:12.581 killing process with pid 3615518 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 3615518 00:05:12.581 19:01:25 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 3615518 00:05:12.850 19:01:25 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:05:12.850 19:01:25 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:05:12.850 19:01:25 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 3615232 ]] 00:05:12.850 19:01:25 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 3615232 00:05:12.850 19:01:25 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 3615232 ']' 00:05:12.850 19:01:25 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 3615232 00:05:12.850 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (3615232) - No such process 00:05:12.850 19:01:25 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 3615232 is not found' 00:05:12.851 Process with pid 3615232 is not found 00:05:12.851 19:01:25 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 3615518 ]] 00:05:12.851 19:01:25 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 3615518 00:05:12.851 19:01:25 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 3615518 ']' 00:05:12.851 19:01:25 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 3615518 00:05:12.851 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (3615518) - No such process 00:05:12.851 19:01:25 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 3615518 is not found' 00:05:12.851 Process with pid 3615518 is not found 00:05:12.851 19:01:25 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:05:12.851 00:05:12.851 real 0m15.138s 00:05:12.851 user 0m27.324s 00:05:12.851 sys 0m4.539s 00:05:12.851 19:01:25 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:12.851 19:01:25 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:12.851 ************************************ 00:05:12.851 END TEST cpu_locks 00:05:12.851 ************************************ 00:05:12.851 00:05:12.851 real 0m40.747s 00:05:12.851 user 1m20.826s 00:05:12.851 sys 0m7.848s 00:05:12.851 19:01:25 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:12.851 19:01:25 event -- common/autotest_common.sh@10 -- # set +x 00:05:12.851 ************************************ 00:05:12.851 END TEST event 00:05:12.851 ************************************ 00:05:12.851 19:01:25 -- spdk/autotest.sh@169 -- # run_test thread /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:05:12.851 19:01:25 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:12.851 19:01:25 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:12.851 19:01:25 -- common/autotest_common.sh@10 -- # set +x 00:05:12.851 ************************************ 00:05:12.851 START TEST thread 00:05:12.851 ************************************ 00:05:12.851 19:01:25 thread -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:05:12.851 * Looking for test storage... 00:05:13.111 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1693 -- # lcov --version 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:13.111 19:01:25 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:13.111 19:01:25 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:13.111 19:01:25 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:13.111 19:01:25 thread -- scripts/common.sh@336 -- # IFS=.-: 00:05:13.111 19:01:25 thread -- scripts/common.sh@336 -- # read -ra ver1 00:05:13.111 19:01:25 thread -- scripts/common.sh@337 -- # IFS=.-: 00:05:13.111 19:01:25 thread -- scripts/common.sh@337 -- # read -ra ver2 00:05:13.111 19:01:25 thread -- scripts/common.sh@338 -- # local 'op=<' 00:05:13.111 19:01:25 thread -- scripts/common.sh@340 -- # ver1_l=2 00:05:13.111 19:01:25 thread -- scripts/common.sh@341 -- # ver2_l=1 00:05:13.111 19:01:25 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:13.111 19:01:25 thread -- scripts/common.sh@344 -- # case "$op" in 00:05:13.111 19:01:25 thread -- scripts/common.sh@345 -- # : 1 00:05:13.111 19:01:25 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:13.111 19:01:25 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:13.111 19:01:25 thread -- scripts/common.sh@365 -- # decimal 1 00:05:13.111 19:01:25 thread -- scripts/common.sh@353 -- # local d=1 00:05:13.111 19:01:25 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:13.111 19:01:25 thread -- scripts/common.sh@355 -- # echo 1 00:05:13.111 19:01:25 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:05:13.111 19:01:25 thread -- scripts/common.sh@366 -- # decimal 2 00:05:13.111 19:01:25 thread -- scripts/common.sh@353 -- # local d=2 00:05:13.111 19:01:25 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:13.111 19:01:25 thread -- scripts/common.sh@355 -- # echo 2 00:05:13.111 19:01:25 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:05:13.111 19:01:25 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:13.111 19:01:25 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:13.111 19:01:25 thread -- scripts/common.sh@368 -- # return 0 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:13.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.111 --rc genhtml_branch_coverage=1 00:05:13.111 --rc genhtml_function_coverage=1 00:05:13.111 --rc genhtml_legend=1 00:05:13.111 --rc geninfo_all_blocks=1 00:05:13.111 --rc geninfo_unexecuted_blocks=1 00:05:13.111 00:05:13.111 ' 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:13.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.111 --rc genhtml_branch_coverage=1 00:05:13.111 --rc genhtml_function_coverage=1 00:05:13.111 --rc genhtml_legend=1 00:05:13.111 --rc geninfo_all_blocks=1 00:05:13.111 --rc geninfo_unexecuted_blocks=1 00:05:13.111 00:05:13.111 ' 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:13.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.111 --rc genhtml_branch_coverage=1 00:05:13.111 --rc genhtml_function_coverage=1 00:05:13.111 --rc genhtml_legend=1 00:05:13.111 --rc geninfo_all_blocks=1 00:05:13.111 --rc geninfo_unexecuted_blocks=1 00:05:13.111 00:05:13.111 ' 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:13.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.111 --rc genhtml_branch_coverage=1 00:05:13.111 --rc genhtml_function_coverage=1 00:05:13.111 --rc genhtml_legend=1 00:05:13.111 --rc geninfo_all_blocks=1 00:05:13.111 --rc geninfo_unexecuted_blocks=1 00:05:13.111 00:05:13.111 ' 00:05:13.111 19:01:25 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:13.111 19:01:25 thread -- common/autotest_common.sh@10 -- # set +x 00:05:13.111 ************************************ 00:05:13.111 START TEST thread_poller_perf 00:05:13.112 ************************************ 00:05:13.112 19:01:25 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:05:13.112 [2024-11-26 19:01:25.636851] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:13.112 [2024-11-26 19:01:25.636937] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3615971 ] 00:05:13.112 [2024-11-26 19:01:25.718187] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:13.371 [2024-11-26 19:01:25.754149] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:13.371 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:05:14.311 [2024-11-26T18:01:26.936Z] ====================================== 00:05:14.311 [2024-11-26T18:01:26.936Z] busy:2407227984 (cyc) 00:05:14.311 [2024-11-26T18:01:26.936Z] total_run_count: 287000 00:05:14.311 [2024-11-26T18:01:26.936Z] tsc_hz: 2400000000 (cyc) 00:05:14.311 [2024-11-26T18:01:26.936Z] ====================================== 00:05:14.311 [2024-11-26T18:01:26.936Z] poller_cost: 8387 (cyc), 3494 (nsec) 00:05:14.311 00:05:14.311 real 0m1.179s 00:05:14.311 user 0m1.107s 00:05:14.311 sys 0m0.067s 00:05:14.311 19:01:26 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:14.311 19:01:26 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:05:14.311 ************************************ 00:05:14.311 END TEST thread_poller_perf 00:05:14.311 ************************************ 00:05:14.311 19:01:26 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:05:14.311 19:01:26 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:05:14.311 19:01:26 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:14.311 19:01:26 thread -- common/autotest_common.sh@10 -- # set +x 00:05:14.311 ************************************ 00:05:14.311 START TEST thread_poller_perf 00:05:14.311 ************************************ 00:05:14.311 19:01:26 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:05:14.311 [2024-11-26 19:01:26.890912] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:14.311 [2024-11-26 19:01:26.891017] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3616321 ] 00:05:14.571 [2024-11-26 19:01:26.971137] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:14.571 [2024-11-26 19:01:27.006275] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:14.571 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:05:15.512 [2024-11-26T18:01:28.137Z] ====================================== 00:05:15.512 [2024-11-26T18:01:28.137Z] busy:2401891386 (cyc) 00:05:15.512 [2024-11-26T18:01:28.137Z] total_run_count: 3812000 00:05:15.512 [2024-11-26T18:01:28.137Z] tsc_hz: 2400000000 (cyc) 00:05:15.512 [2024-11-26T18:01:28.137Z] ====================================== 00:05:15.512 [2024-11-26T18:01:28.137Z] poller_cost: 630 (cyc), 262 (nsec) 00:05:15.512 00:05:15.512 real 0m1.169s 00:05:15.512 user 0m1.098s 00:05:15.512 sys 0m0.067s 00:05:15.512 19:01:28 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:15.512 19:01:28 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:05:15.512 ************************************ 00:05:15.512 END TEST thread_poller_perf 00:05:15.512 ************************************ 00:05:15.512 19:01:28 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:05:15.512 00:05:15.512 real 0m2.694s 00:05:15.512 user 0m2.384s 00:05:15.512 sys 0m0.321s 00:05:15.512 19:01:28 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:15.512 19:01:28 thread -- common/autotest_common.sh@10 -- # set +x 00:05:15.512 ************************************ 00:05:15.512 END TEST thread 00:05:15.512 ************************************ 00:05:15.512 19:01:28 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:05:15.512 19:01:28 -- spdk/autotest.sh@176 -- # run_test app_cmdline /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:05:15.512 19:01:28 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:15.512 19:01:28 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:15.512 19:01:28 -- common/autotest_common.sh@10 -- # set +x 00:05:15.773 ************************************ 00:05:15.773 START TEST app_cmdline 00:05:15.773 ************************************ 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:05:15.773 * Looking for test storage... 00:05:15.773 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1693 -- # lcov --version 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@345 -- # : 1 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:15.773 19:01:28 app_cmdline -- scripts/common.sh@368 -- # return 0 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:15.773 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.773 --rc genhtml_branch_coverage=1 00:05:15.773 --rc genhtml_function_coverage=1 00:05:15.773 --rc genhtml_legend=1 00:05:15.773 --rc geninfo_all_blocks=1 00:05:15.773 --rc geninfo_unexecuted_blocks=1 00:05:15.773 00:05:15.773 ' 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:15.773 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.773 --rc genhtml_branch_coverage=1 00:05:15.773 --rc genhtml_function_coverage=1 00:05:15.773 --rc genhtml_legend=1 00:05:15.773 --rc geninfo_all_blocks=1 00:05:15.773 --rc geninfo_unexecuted_blocks=1 00:05:15.773 00:05:15.773 ' 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:15.773 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.773 --rc genhtml_branch_coverage=1 00:05:15.773 --rc genhtml_function_coverage=1 00:05:15.773 --rc genhtml_legend=1 00:05:15.773 --rc geninfo_all_blocks=1 00:05:15.773 --rc geninfo_unexecuted_blocks=1 00:05:15.773 00:05:15.773 ' 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:15.773 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.773 --rc genhtml_branch_coverage=1 00:05:15.773 --rc genhtml_function_coverage=1 00:05:15.773 --rc genhtml_legend=1 00:05:15.773 --rc geninfo_all_blocks=1 00:05:15.773 --rc geninfo_unexecuted_blocks=1 00:05:15.773 00:05:15.773 ' 00:05:15.773 19:01:28 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:05:15.773 19:01:28 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=3616725 00:05:15.773 19:01:28 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 3616725 00:05:15.773 19:01:28 app_cmdline -- app/cmdline.sh@16 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 3616725 ']' 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:15.773 19:01:28 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:15.774 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:15.774 19:01:28 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:15.774 19:01:28 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:05:16.034 [2024-11-26 19:01:28.412251] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:16.034 [2024-11-26 19:01:28.412302] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3616725 ] 00:05:16.034 [2024-11-26 19:01:28.490370] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:16.034 [2024-11-26 19:01:28.527137] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:16.604 19:01:29 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:16.604 19:01:29 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:05:16.604 19:01:29 app_cmdline -- app/cmdline.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py spdk_get_version 00:05:16.864 { 00:05:16.864 "version": "SPDK v25.01-pre git sha1 bb877d8c1", 00:05:16.864 "fields": { 00:05:16.864 "major": 25, 00:05:16.864 "minor": 1, 00:05:16.864 "patch": 0, 00:05:16.864 "suffix": "-pre", 00:05:16.864 "commit": "bb877d8c1" 00:05:16.864 } 00:05:16.864 } 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@26 -- # sort 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:05:16.864 19:01:29 app_cmdline -- app/cmdline.sh@30 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:05:16.864 19:01:29 app_cmdline -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:05:17.125 request: 00:05:17.125 { 00:05:17.125 "method": "env_dpdk_get_mem_stats", 00:05:17.125 "req_id": 1 00:05:17.125 } 00:05:17.125 Got JSON-RPC error response 00:05:17.125 response: 00:05:17.125 { 00:05:17.125 "code": -32601, 00:05:17.125 "message": "Method not found" 00:05:17.125 } 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:17.125 19:01:29 app_cmdline -- app/cmdline.sh@1 -- # killprocess 3616725 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 3616725 ']' 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 3616725 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3616725 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3616725' 00:05:17.125 killing process with pid 3616725 00:05:17.125 19:01:29 app_cmdline -- common/autotest_common.sh@973 -- # kill 3616725 00:05:17.126 19:01:29 app_cmdline -- common/autotest_common.sh@978 -- # wait 3616725 00:05:17.387 00:05:17.387 real 0m1.735s 00:05:17.387 user 0m2.089s 00:05:17.387 sys 0m0.456s 00:05:17.387 19:01:29 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:17.387 19:01:29 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:05:17.387 ************************************ 00:05:17.387 END TEST app_cmdline 00:05:17.387 ************************************ 00:05:17.387 19:01:29 -- spdk/autotest.sh@177 -- # run_test version /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:05:17.387 19:01:29 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:17.387 19:01:29 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:17.387 19:01:29 -- common/autotest_common.sh@10 -- # set +x 00:05:17.387 ************************************ 00:05:17.387 START TEST version 00:05:17.387 ************************************ 00:05:17.387 19:01:29 version -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:05:17.650 * Looking for test storage... 00:05:17.650 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:05:17.650 19:01:30 version -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:17.650 19:01:30 version -- common/autotest_common.sh@1693 -- # lcov --version 00:05:17.650 19:01:30 version -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:17.650 19:01:30 version -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:17.650 19:01:30 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:17.650 19:01:30 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:17.650 19:01:30 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:17.650 19:01:30 version -- scripts/common.sh@336 -- # IFS=.-: 00:05:17.650 19:01:30 version -- scripts/common.sh@336 -- # read -ra ver1 00:05:17.650 19:01:30 version -- scripts/common.sh@337 -- # IFS=.-: 00:05:17.650 19:01:30 version -- scripts/common.sh@337 -- # read -ra ver2 00:05:17.650 19:01:30 version -- scripts/common.sh@338 -- # local 'op=<' 00:05:17.650 19:01:30 version -- scripts/common.sh@340 -- # ver1_l=2 00:05:17.650 19:01:30 version -- scripts/common.sh@341 -- # ver2_l=1 00:05:17.650 19:01:30 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:17.650 19:01:30 version -- scripts/common.sh@344 -- # case "$op" in 00:05:17.650 19:01:30 version -- scripts/common.sh@345 -- # : 1 00:05:17.650 19:01:30 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:17.650 19:01:30 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:17.650 19:01:30 version -- scripts/common.sh@365 -- # decimal 1 00:05:17.650 19:01:30 version -- scripts/common.sh@353 -- # local d=1 00:05:17.650 19:01:30 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:17.650 19:01:30 version -- scripts/common.sh@355 -- # echo 1 00:05:17.650 19:01:30 version -- scripts/common.sh@365 -- # ver1[v]=1 00:05:17.650 19:01:30 version -- scripts/common.sh@366 -- # decimal 2 00:05:17.650 19:01:30 version -- scripts/common.sh@353 -- # local d=2 00:05:17.651 19:01:30 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:17.651 19:01:30 version -- scripts/common.sh@355 -- # echo 2 00:05:17.651 19:01:30 version -- scripts/common.sh@366 -- # ver2[v]=2 00:05:17.651 19:01:30 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:17.651 19:01:30 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:17.651 19:01:30 version -- scripts/common.sh@368 -- # return 0 00:05:17.651 19:01:30 version -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:17.651 19:01:30 version -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:17.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.651 --rc genhtml_branch_coverage=1 00:05:17.651 --rc genhtml_function_coverage=1 00:05:17.651 --rc genhtml_legend=1 00:05:17.651 --rc geninfo_all_blocks=1 00:05:17.651 --rc geninfo_unexecuted_blocks=1 00:05:17.651 00:05:17.651 ' 00:05:17.651 19:01:30 version -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:17.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.651 --rc genhtml_branch_coverage=1 00:05:17.651 --rc genhtml_function_coverage=1 00:05:17.651 --rc genhtml_legend=1 00:05:17.651 --rc geninfo_all_blocks=1 00:05:17.651 --rc geninfo_unexecuted_blocks=1 00:05:17.651 00:05:17.651 ' 00:05:17.651 19:01:30 version -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:17.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.651 --rc genhtml_branch_coverage=1 00:05:17.651 --rc genhtml_function_coverage=1 00:05:17.651 --rc genhtml_legend=1 00:05:17.651 --rc geninfo_all_blocks=1 00:05:17.651 --rc geninfo_unexecuted_blocks=1 00:05:17.651 00:05:17.651 ' 00:05:17.651 19:01:30 version -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:17.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.651 --rc genhtml_branch_coverage=1 00:05:17.651 --rc genhtml_function_coverage=1 00:05:17.651 --rc genhtml_legend=1 00:05:17.651 --rc geninfo_all_blocks=1 00:05:17.651 --rc geninfo_unexecuted_blocks=1 00:05:17.651 00:05:17.651 ' 00:05:17.651 19:01:30 version -- app/version.sh@17 -- # get_header_version major 00:05:17.651 19:01:30 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # cut -f2 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # tr -d '"' 00:05:17.651 19:01:30 version -- app/version.sh@17 -- # major=25 00:05:17.651 19:01:30 version -- app/version.sh@18 -- # get_header_version minor 00:05:17.651 19:01:30 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # cut -f2 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # tr -d '"' 00:05:17.651 19:01:30 version -- app/version.sh@18 -- # minor=1 00:05:17.651 19:01:30 version -- app/version.sh@19 -- # get_header_version patch 00:05:17.651 19:01:30 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # cut -f2 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # tr -d '"' 00:05:17.651 19:01:30 version -- app/version.sh@19 -- # patch=0 00:05:17.651 19:01:30 version -- app/version.sh@20 -- # get_header_version suffix 00:05:17.651 19:01:30 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # cut -f2 00:05:17.651 19:01:30 version -- app/version.sh@14 -- # tr -d '"' 00:05:17.651 19:01:30 version -- app/version.sh@20 -- # suffix=-pre 00:05:17.651 19:01:30 version -- app/version.sh@22 -- # version=25.1 00:05:17.651 19:01:30 version -- app/version.sh@25 -- # (( patch != 0 )) 00:05:17.651 19:01:30 version -- app/version.sh@28 -- # version=25.1rc0 00:05:17.651 19:01:30 version -- app/version.sh@30 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:05:17.651 19:01:30 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:05:17.651 19:01:30 version -- app/version.sh@30 -- # py_version=25.1rc0 00:05:17.651 19:01:30 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:05:17.651 00:05:17.651 real 0m0.265s 00:05:17.651 user 0m0.157s 00:05:17.651 sys 0m0.158s 00:05:17.651 19:01:30 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:17.651 19:01:30 version -- common/autotest_common.sh@10 -- # set +x 00:05:17.651 ************************************ 00:05:17.651 END TEST version 00:05:17.651 ************************************ 00:05:17.912 19:01:30 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:05:17.912 19:01:30 -- spdk/autotest.sh@188 -- # [[ 0 -eq 1 ]] 00:05:17.912 19:01:30 -- spdk/autotest.sh@194 -- # uname -s 00:05:17.912 19:01:30 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:05:17.912 19:01:30 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:05:17.912 19:01:30 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:05:17.912 19:01:30 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:05:17.912 19:01:30 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:05:17.912 19:01:30 -- spdk/autotest.sh@260 -- # timing_exit lib 00:05:17.912 19:01:30 -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:17.912 19:01:30 -- common/autotest_common.sh@10 -- # set +x 00:05:17.912 19:01:30 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:05:17.912 19:01:30 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:05:17.913 19:01:30 -- spdk/autotest.sh@276 -- # '[' 1 -eq 1 ']' 00:05:17.913 19:01:30 -- spdk/autotest.sh@277 -- # export NET_TYPE 00:05:17.913 19:01:30 -- spdk/autotest.sh@280 -- # '[' tcp = rdma ']' 00:05:17.913 19:01:30 -- spdk/autotest.sh@283 -- # '[' tcp = tcp ']' 00:05:17.913 19:01:30 -- spdk/autotest.sh@284 -- # run_test nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:05:17.913 19:01:30 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:05:17.913 19:01:30 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:17.913 19:01:30 -- common/autotest_common.sh@10 -- # set +x 00:05:17.913 ************************************ 00:05:17.913 START TEST nvmf_tcp 00:05:17.913 ************************************ 00:05:17.913 19:01:30 nvmf_tcp -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:05:17.913 * Looking for test storage... 00:05:17.913 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:05:17.913 19:01:30 nvmf_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:17.913 19:01:30 nvmf_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:05:17.913 19:01:30 nvmf_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:18.174 19:01:30 nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:18.174 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.174 --rc genhtml_branch_coverage=1 00:05:18.174 --rc genhtml_function_coverage=1 00:05:18.174 --rc genhtml_legend=1 00:05:18.174 --rc geninfo_all_blocks=1 00:05:18.174 --rc geninfo_unexecuted_blocks=1 00:05:18.174 00:05:18.174 ' 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:18.174 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.174 --rc genhtml_branch_coverage=1 00:05:18.174 --rc genhtml_function_coverage=1 00:05:18.174 --rc genhtml_legend=1 00:05:18.174 --rc geninfo_all_blocks=1 00:05:18.174 --rc geninfo_unexecuted_blocks=1 00:05:18.174 00:05:18.174 ' 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:18.174 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.174 --rc genhtml_branch_coverage=1 00:05:18.174 --rc genhtml_function_coverage=1 00:05:18.174 --rc genhtml_legend=1 00:05:18.174 --rc geninfo_all_blocks=1 00:05:18.174 --rc geninfo_unexecuted_blocks=1 00:05:18.174 00:05:18.174 ' 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:18.174 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.174 --rc genhtml_branch_coverage=1 00:05:18.174 --rc genhtml_function_coverage=1 00:05:18.174 --rc genhtml_legend=1 00:05:18.174 --rc geninfo_all_blocks=1 00:05:18.174 --rc geninfo_unexecuted_blocks=1 00:05:18.174 00:05:18.174 ' 00:05:18.174 19:01:30 nvmf_tcp -- nvmf/nvmf.sh@10 -- # uname -s 00:05:18.174 19:01:30 nvmf_tcp -- nvmf/nvmf.sh@10 -- # '[' '!' Linux = Linux ']' 00:05:18.174 19:01:30 nvmf_tcp -- nvmf/nvmf.sh@14 -- # run_test nvmf_target_core /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:18.174 19:01:30 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:18.174 ************************************ 00:05:18.174 START TEST nvmf_target_core 00:05:18.174 ************************************ 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:05:18.174 * Looking for test storage... 00:05:18.174 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1693 -- # lcov --version 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # IFS=.-: 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # read -ra ver1 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # IFS=.-: 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # read -ra ver2 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@338 -- # local 'op=<' 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@340 -- # ver1_l=2 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@341 -- # ver2_l=1 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@344 -- # case "$op" in 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@345 -- # : 1 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # decimal 1 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=1 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 1 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # ver1[v]=1 00:05:18.174 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # decimal 2 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=2 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 2 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # ver2[v]=2 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # return 0 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:18.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.436 --rc genhtml_branch_coverage=1 00:05:18.436 --rc genhtml_function_coverage=1 00:05:18.436 --rc genhtml_legend=1 00:05:18.436 --rc geninfo_all_blocks=1 00:05:18.436 --rc geninfo_unexecuted_blocks=1 00:05:18.436 00:05:18.436 ' 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:18.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.436 --rc genhtml_branch_coverage=1 00:05:18.436 --rc genhtml_function_coverage=1 00:05:18.436 --rc genhtml_legend=1 00:05:18.436 --rc geninfo_all_blocks=1 00:05:18.436 --rc geninfo_unexecuted_blocks=1 00:05:18.436 00:05:18.436 ' 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:18.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.436 --rc genhtml_branch_coverage=1 00:05:18.436 --rc genhtml_function_coverage=1 00:05:18.436 --rc genhtml_legend=1 00:05:18.436 --rc geninfo_all_blocks=1 00:05:18.436 --rc geninfo_unexecuted_blocks=1 00:05:18.436 00:05:18.436 ' 00:05:18.436 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:18.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.436 --rc genhtml_branch_coverage=1 00:05:18.437 --rc genhtml_function_coverage=1 00:05:18.437 --rc genhtml_legend=1 00:05:18.437 --rc geninfo_all_blocks=1 00:05:18.437 --rc geninfo_unexecuted_blocks=1 00:05:18.437 00:05:18.437 ' 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # uname -s 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@15 -- # shopt -s extglob 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- paths/export.sh@5 -- # export PATH 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@51 -- # : 0 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:18.437 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:05:18.437 ************************************ 00:05:18.437 START TEST nvmf_abort 00:05:18.437 ************************************ 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:05:18.437 * Looking for test storage... 00:05:18.437 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1693 -- # lcov --version 00:05:18.437 19:01:30 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:18.700 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.700 --rc genhtml_branch_coverage=1 00:05:18.700 --rc genhtml_function_coverage=1 00:05:18.700 --rc genhtml_legend=1 00:05:18.700 --rc geninfo_all_blocks=1 00:05:18.700 --rc geninfo_unexecuted_blocks=1 00:05:18.700 00:05:18.700 ' 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:18.700 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.700 --rc genhtml_branch_coverage=1 00:05:18.700 --rc genhtml_function_coverage=1 00:05:18.700 --rc genhtml_legend=1 00:05:18.700 --rc geninfo_all_blocks=1 00:05:18.700 --rc geninfo_unexecuted_blocks=1 00:05:18.700 00:05:18.700 ' 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:18.700 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.700 --rc genhtml_branch_coverage=1 00:05:18.700 --rc genhtml_function_coverage=1 00:05:18.700 --rc genhtml_legend=1 00:05:18.700 --rc geninfo_all_blocks=1 00:05:18.700 --rc geninfo_unexecuted_blocks=1 00:05:18.700 00:05:18.700 ' 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:18.700 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:18.700 --rc genhtml_branch_coverage=1 00:05:18.700 --rc genhtml_function_coverage=1 00:05:18.700 --rc genhtml_legend=1 00:05:18.700 --rc geninfo_all_blocks=1 00:05:18.700 --rc geninfo_unexecuted_blocks=1 00:05:18.700 00:05:18.700 ' 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:18.700 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:18.701 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@476 -- # prepare_net_devs 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@438 -- # local -g is_hw=no 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # remove_spdk_ns 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:05:18.701 19:01:31 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:05:26.847 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:05:26.848 Found 0000:31:00.0 (0x8086 - 0x159b) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:05:26.848 Found 0000:31:00.1 (0x8086 - 0x159b) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:05:26.848 Found net devices under 0000:31:00.0: cvl_0_0 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:05:26.848 Found net devices under 0000:31:00.1: cvl_0_1 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # is_hw=yes 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:05:26.848 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:05:27.108 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:05:27.108 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:05:27.108 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:05:27.108 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:05:27.108 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:05:27.108 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:05:27.108 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:05:27.108 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:05:27.108 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.642 ms 00:05:27.108 00:05:27.108 --- 10.0.0.2 ping statistics --- 00:05:27.109 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:05:27.109 rtt min/avg/max/mdev = 0.642/0.642/0.642/0.000 ms 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:05:27.109 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:05:27.109 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.275 ms 00:05:27.109 00:05:27.109 --- 10.0.0.1 ping statistics --- 00:05:27.109 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:05:27.109 rtt min/avg/max/mdev = 0.275/0.275/0.275/0.000 ms 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@450 -- # return 0 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@509 -- # nvmfpid=3621595 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@510 -- # waitforlisten 3621595 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@835 -- # '[' -z 3621595 ']' 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:27.109 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:27.109 19:01:39 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:27.368 [2024-11-26 19:01:39.736414] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:27.368 [2024-11-26 19:01:39.736468] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:05:27.368 [2024-11-26 19:01:39.843267] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:27.368 [2024-11-26 19:01:39.897114] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:05:27.368 [2024-11-26 19:01:39.897166] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:05:27.368 [2024-11-26 19:01:39.897174] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:27.368 [2024-11-26 19:01:39.897181] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:27.368 [2024-11-26 19:01:39.897188] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:05:27.368 [2024-11-26 19:01:39.899236] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:27.368 [2024-11-26 19:01:39.899402] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:27.368 [2024-11-26 19:01:39.899402] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:05:27.978 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:27.978 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@868 -- # return 0 00:05:27.978 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:05:27.978 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:27.978 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 [2024-11-26 19:01:40.591671] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 Malloc0 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 Delay0 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 [2024-11-26 19:01:40.671078] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.241 19:01:40 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:05:28.241 [2024-11-26 19:01:40.800278] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:05:30.791 Initializing NVMe Controllers 00:05:30.791 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:05:30.791 controller IO queue size 128 less than required 00:05:30.791 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:05:30.791 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:05:30.791 Initialization complete. Launching workers. 00:05:30.791 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 28973 00:05:30.791 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 29034, failed to submit 62 00:05:30.791 success 28977, unsuccessful 57, failed 0 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@516 -- # nvmfcleanup 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:05:30.791 rmmod nvme_tcp 00:05:30.791 rmmod nvme_fabrics 00:05:30.791 rmmod nvme_keyring 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@517 -- # '[' -n 3621595 ']' 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@518 -- # killprocess 3621595 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@954 -- # '[' -z 3621595 ']' 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@958 -- # kill -0 3621595 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@959 -- # uname 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:30.791 19:01:42 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3621595 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3621595' 00:05:30.791 killing process with pid 3621595 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@973 -- # kill 3621595 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@978 -- # wait 3621595 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@791 -- # iptables-save 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@791 -- # iptables-restore 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:05:30.791 19:01:43 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:05:32.704 00:05:32.704 real 0m14.347s 00:05:32.704 user 0m14.010s 00:05:32.704 sys 0m7.319s 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:05:32.704 ************************************ 00:05:32.704 END TEST nvmf_abort 00:05:32.704 ************************************ 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:05:32.704 ************************************ 00:05:32.704 START TEST nvmf_ns_hotplug_stress 00:05:32.704 ************************************ 00:05:32.704 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:05:32.967 * Looking for test storage... 00:05:32.967 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1693 -- # lcov --version 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:32.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:32.967 --rc genhtml_branch_coverage=1 00:05:32.967 --rc genhtml_function_coverage=1 00:05:32.967 --rc genhtml_legend=1 00:05:32.967 --rc geninfo_all_blocks=1 00:05:32.967 --rc geninfo_unexecuted_blocks=1 00:05:32.967 00:05:32.967 ' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:32.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:32.967 --rc genhtml_branch_coverage=1 00:05:32.967 --rc genhtml_function_coverage=1 00:05:32.967 --rc genhtml_legend=1 00:05:32.967 --rc geninfo_all_blocks=1 00:05:32.967 --rc geninfo_unexecuted_blocks=1 00:05:32.967 00:05:32.967 ' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:32.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:32.967 --rc genhtml_branch_coverage=1 00:05:32.967 --rc genhtml_function_coverage=1 00:05:32.967 --rc genhtml_legend=1 00:05:32.967 --rc geninfo_all_blocks=1 00:05:32.967 --rc geninfo_unexecuted_blocks=1 00:05:32.967 00:05:32.967 ' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:32.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:32.967 --rc genhtml_branch_coverage=1 00:05:32.967 --rc genhtml_function_coverage=1 00:05:32.967 --rc genhtml_legend=1 00:05:32.967 --rc geninfo_all_blocks=1 00:05:32.967 --rc geninfo_unexecuted_blocks=1 00:05:32.967 00:05:32.967 ' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:05:32.967 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:32.968 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # prepare_net_devs 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # local -g is_hw=no 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # remove_spdk_ns 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:05:32.968 19:01:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:05:41.117 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:05:41.118 Found 0000:31:00.0 (0x8086 - 0x159b) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:05:41.118 Found 0000:31:00.1 (0x8086 - 0x159b) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:05:41.118 Found net devices under 0000:31:00.0: cvl_0_0 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:05:41.118 Found net devices under 0000:31:00.1: cvl_0_1 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # is_hw=yes 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:05:41.118 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:05:41.119 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:05:41.119 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:05:41.119 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:05:41.119 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:05:41.380 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:05:41.380 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:05:41.381 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:05:41.381 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.640 ms 00:05:41.381 00:05:41.381 --- 10.0.0.2 ping statistics --- 00:05:41.381 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:05:41.381 rtt min/avg/max/mdev = 0.640/0.640/0.640/0.000 ms 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:05:41.381 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:05:41.381 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.319 ms 00:05:41.381 00:05:41.381 --- 10.0.0.1 ping statistics --- 00:05:41.381 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:05:41.381 rtt min/avg/max/mdev = 0.319/0.319/0.319/0.000 ms 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@450 -- # return 0 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # nvmfpid=3627084 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # waitforlisten 3627084 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # '[' -z 3627084 ']' 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:41.381 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:41.381 19:01:53 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:05:41.381 [2024-11-26 19:01:53.910003] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:05:41.381 [2024-11-26 19:01:53.910064] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:05:41.643 [2024-11-26 19:01:54.021788] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:41.643 [2024-11-26 19:01:54.073317] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:05:41.643 [2024-11-26 19:01:54.073373] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:05:41.643 [2024-11-26 19:01:54.073382] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:41.643 [2024-11-26 19:01:54.073389] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:41.643 [2024-11-26 19:01:54.073395] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:05:41.643 [2024-11-26 19:01:54.075248] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:41.643 [2024-11-26 19:01:54.075414] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:41.643 [2024-11-26 19:01:54.075414] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@868 -- # return 0 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:05:42.215 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:05:42.475 [2024-11-26 19:01:54.917289] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:42.475 19:01:54 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:05:42.737 19:01:55 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:05:42.737 [2024-11-26 19:01:55.278735] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:05:42.737 19:01:55 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:05:42.998 19:01:55 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:05:43.259 Malloc0 00:05:43.259 19:01:55 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:05:43.259 Delay0 00:05:43.259 19:01:55 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:43.521 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:05:43.782 NULL1 00:05:43.782 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:05:44.043 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=3627686 00:05:44.043 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:05:44.043 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:44.043 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:44.043 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:44.334 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:05:44.334 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:05:44.334 true 00:05:44.594 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:44.594 19:01:56 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:44.594 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:44.856 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:05:44.856 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:05:45.117 true 00:05:45.117 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:45.117 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:45.117 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:45.379 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:05:45.379 19:01:57 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:05:45.640 true 00:05:45.640 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:45.640 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:45.640 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:45.901 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:05:45.901 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:05:46.162 true 00:05:46.162 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:46.162 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:46.423 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:46.423 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:05:46.423 19:01:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:05:46.684 true 00:05:46.685 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:46.685 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:46.946 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:46.946 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:05:46.946 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:05:47.207 true 00:05:47.207 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:47.207 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:47.468 19:01:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:47.468 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:05:47.468 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:05:47.730 true 00:05:47.730 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:47.730 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:47.992 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:48.254 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:05:48.254 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:05:48.254 true 00:05:48.254 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:48.254 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:48.515 19:02:00 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:48.776 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:05:48.776 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:05:48.776 true 00:05:48.776 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:48.776 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:49.037 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:49.298 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:05:49.298 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:05:49.298 true 00:05:49.298 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:49.298 19:02:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:49.559 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:49.820 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:05:49.820 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:05:49.820 true 00:05:50.082 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:50.082 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:50.082 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:50.344 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:05:50.344 19:02:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:05:50.605 true 00:05:50.605 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:50.605 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:50.605 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:50.867 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:05:50.867 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:05:51.128 true 00:05:51.128 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:51.128 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:51.128 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:51.389 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:05:51.389 19:02:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:05:51.650 true 00:05:51.650 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:51.650 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:51.650 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:51.911 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:05:51.911 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:05:52.172 true 00:05:52.172 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:52.173 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:52.434 19:02:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:52.434 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:05:52.434 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:05:52.694 true 00:05:52.694 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:52.694 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:52.954 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:52.954 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:05:52.954 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:05:53.217 true 00:05:53.217 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:53.217 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:53.479 19:02:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:53.479 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:05:53.479 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:05:53.741 true 00:05:53.741 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:53.741 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:54.001 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:54.262 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:05:54.262 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:05:54.262 true 00:05:54.262 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:54.262 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:54.522 19:02:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:54.522 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:05:54.523 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:05:54.784 true 00:05:54.784 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:54.784 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:55.044 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:55.306 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:05:55.306 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:05:55.306 true 00:05:55.306 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:55.306 19:02:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:55.566 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:55.827 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:05:55.827 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:05:55.827 true 00:05:55.827 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:55.827 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:56.087 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:56.347 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:05:56.347 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:05:56.347 true 00:05:56.347 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:56.347 19:02:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:56.609 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:56.869 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:05:56.869 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:05:56.869 true 00:05:57.130 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:57.130 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:57.130 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:57.391 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:05:57.391 19:02:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:05:57.391 true 00:05:57.651 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:57.651 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:57.651 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:57.911 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:05:57.911 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:05:58.171 true 00:05:58.171 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:58.171 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:58.171 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:58.431 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:05:58.431 19:02:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:05:58.693 true 00:05:58.693 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:58.693 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:58.693 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:58.953 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:05:58.953 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:05:59.214 true 00:05:59.214 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:59.214 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:59.214 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:59.475 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:05:59.475 19:02:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:05:59.736 true 00:05:59.736 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:05:59.736 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:59.736 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:59.997 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1030 00:05:59.997 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1030 00:06:00.297 true 00:06:00.297 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:00.297 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:00.297 19:02:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:00.615 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1031 00:06:00.615 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1031 00:06:00.615 true 00:06:00.903 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:00.903 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:00.903 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:01.164 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1032 00:06:01.164 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1032 00:06:01.164 true 00:06:01.164 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:01.164 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:01.426 19:02:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:01.687 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1033 00:06:01.687 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1033 00:06:01.947 true 00:06:01.947 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:01.947 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:01.947 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:02.207 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1034 00:06:02.207 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1034 00:06:02.467 true 00:06:02.467 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:02.467 19:02:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:02.727 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:02.727 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1035 00:06:02.727 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1035 00:06:02.988 true 00:06:02.989 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:02.989 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:03.250 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:03.250 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1036 00:06:03.250 19:02:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1036 00:06:03.510 true 00:06:03.510 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:03.510 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:03.769 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:04.029 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1037 00:06:04.029 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1037 00:06:04.029 true 00:06:04.029 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:04.029 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:04.289 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:04.549 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1038 00:06:04.549 19:02:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1038 00:06:04.549 true 00:06:04.550 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:04.550 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:04.812 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:05.073 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1039 00:06:05.073 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1039 00:06:05.073 true 00:06:05.073 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:05.073 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:05.335 19:02:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:05.595 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1040 00:06:05.595 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1040 00:06:05.595 true 00:06:05.855 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:05.855 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:05.855 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:06.115 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1041 00:06:06.115 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1041 00:06:06.375 true 00:06:06.375 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:06.375 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:06.375 19:02:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:06.635 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1042 00:06:06.635 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1042 00:06:06.896 true 00:06:06.896 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:06.896 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:06.896 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:07.155 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1043 00:06:07.155 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1043 00:06:07.441 true 00:06:07.441 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:07.441 19:02:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:07.441 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:07.701 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1044 00:06:07.701 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1044 00:06:07.961 true 00:06:07.961 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:07.961 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:08.222 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:08.222 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1045 00:06:08.222 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1045 00:06:08.482 true 00:06:08.482 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:08.482 19:02:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:08.741 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:08.741 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1046 00:06:08.741 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1046 00:06:09.001 true 00:06:09.001 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:09.001 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:09.262 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:09.262 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1047 00:06:09.262 19:02:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1047 00:06:09.522 true 00:06:09.522 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:09.522 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:09.783 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:09.783 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1048 00:06:09.783 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1048 00:06:10.044 true 00:06:10.044 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:10.044 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:10.305 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:10.565 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1049 00:06:10.565 19:02:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1049 00:06:10.565 true 00:06:10.565 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:10.565 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:10.825 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:11.085 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1050 00:06:11.085 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1050 00:06:11.085 true 00:06:11.085 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:11.085 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:11.346 19:02:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:11.606 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1051 00:06:11.606 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1051 00:06:11.606 true 00:06:11.606 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:11.606 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:11.867 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:12.127 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1052 00:06:12.127 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1052 00:06:12.127 true 00:06:12.127 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:12.127 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:12.388 19:02:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:12.649 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1053 00:06:12.649 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1053 00:06:12.649 true 00:06:12.909 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:12.909 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:12.909 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:13.169 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1054 00:06:13.169 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1054 00:06:13.431 true 00:06:13.431 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:13.431 19:02:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:13.431 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:13.692 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1055 00:06:13.692 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1055 00:06:13.953 true 00:06:13.953 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:13.953 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:13.953 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:14.214 Initializing NVMe Controllers 00:06:14.214 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:06:14.214 Controller IO queue size 128, less than required. 00:06:14.214 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:06:14.214 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:06:14.214 Initialization complete. Launching workers. 00:06:14.214 ======================================================== 00:06:14.214 Latency(us) 00:06:14.214 Device Information : IOPS MiB/s Average min max 00:06:14.214 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 30205.73 14.75 4237.43 1425.85 10505.89 00:06:14.214 ======================================================== 00:06:14.214 Total : 30205.73 14.75 4237.43 1425.85 10505.89 00:06:14.214 00:06:14.214 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1056 00:06:14.214 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1056 00:06:14.475 true 00:06:14.475 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 3627686 00:06:14.475 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (3627686) - No such process 00:06:14.475 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 3627686 00:06:14.475 19:02:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:14.735 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:14.735 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:06:14.735 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:06:14.735 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:06:14.735 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:14.735 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:06:14.996 null0 00:06:14.996 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:14.996 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:14.996 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:06:14.996 null1 00:06:15.256 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:15.256 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:15.256 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:06:15.256 null2 00:06:15.256 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:15.256 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:15.256 19:02:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:06:15.517 null3 00:06:15.517 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:15.517 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:15.517 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:06:15.777 null4 00:06:15.777 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:15.777 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:15.777 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:06:15.777 null5 00:06:15.777 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:15.778 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:15.778 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:06:16.038 null6 00:06:16.038 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:16.038 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:16.038 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:06:16.300 null7 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:06:16.300 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 3634243 3634244 3634246 3634248 3634251 3634253 3634254 3634256 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:16.301 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:16.563 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:16.563 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:16.563 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:16.563 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:16.563 19:02:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.563 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:16.824 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:17.085 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.347 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:17.608 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.608 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.608 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:17.608 19:02:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.608 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:17.869 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:17.870 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:18.131 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:18.393 19:02:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:18.393 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.393 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.393 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.656 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:18.919 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:19.181 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.443 19:02:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:06:19.443 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:19.443 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.705 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.706 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:19.970 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # nvmfcleanup 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:06:20.232 rmmod nvme_tcp 00:06:20.232 rmmod nvme_fabrics 00:06:20.232 rmmod nvme_keyring 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@517 -- # '[' -n 3627084 ']' 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # killprocess 3627084 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # '[' -z 3627084 ']' 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@958 -- # kill -0 3627084 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # uname 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3627084 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3627084' 00:06:20.232 killing process with pid 3627084 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@973 -- # kill 3627084 00:06:20.232 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@978 -- # wait 3627084 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-restore 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-save 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:20.494 19:02:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:22.410 19:02:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:06:22.410 00:06:22.410 real 0m49.649s 00:06:22.410 user 3m20.295s 00:06:22.410 sys 0m17.565s 00:06:22.410 19:02:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:22.410 19:02:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:06:22.410 ************************************ 00:06:22.410 END TEST nvmf_ns_hotplug_stress 00:06:22.411 ************************************ 00:06:22.411 19:02:34 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:06:22.411 19:02:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:22.411 19:02:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:22.411 19:02:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:06:22.672 ************************************ 00:06:22.672 START TEST nvmf_delete_subsystem 00:06:22.672 ************************************ 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:06:22.672 * Looking for test storage... 00:06:22.672 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1693 -- # lcov --version 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:22.672 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:22.672 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:22.672 --rc genhtml_branch_coverage=1 00:06:22.672 --rc genhtml_function_coverage=1 00:06:22.672 --rc genhtml_legend=1 00:06:22.672 --rc geninfo_all_blocks=1 00:06:22.672 --rc geninfo_unexecuted_blocks=1 00:06:22.672 00:06:22.672 ' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:22.673 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:22.673 --rc genhtml_branch_coverage=1 00:06:22.673 --rc genhtml_function_coverage=1 00:06:22.673 --rc genhtml_legend=1 00:06:22.673 --rc geninfo_all_blocks=1 00:06:22.673 --rc geninfo_unexecuted_blocks=1 00:06:22.673 00:06:22.673 ' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:22.673 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:22.673 --rc genhtml_branch_coverage=1 00:06:22.673 --rc genhtml_function_coverage=1 00:06:22.673 --rc genhtml_legend=1 00:06:22.673 --rc geninfo_all_blocks=1 00:06:22.673 --rc geninfo_unexecuted_blocks=1 00:06:22.673 00:06:22.673 ' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:22.673 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:22.673 --rc genhtml_branch_coverage=1 00:06:22.673 --rc genhtml_function_coverage=1 00:06:22.673 --rc genhtml_legend=1 00:06:22.673 --rc geninfo_all_blocks=1 00:06:22.673 --rc geninfo_unexecuted_blocks=1 00:06:22.673 00:06:22.673 ' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:22.673 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # prepare_net_devs 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # local -g is_hw=no 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # remove_spdk_ns 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:06:22.673 19:02:35 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:30.812 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:06:30.812 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:06:30.812 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:06:30.812 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:06:30.812 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:06:30.813 Found 0000:31:00.0 (0x8086 - 0x159b) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:06:30.813 Found 0000:31:00.1 (0x8086 - 0x159b) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:06:30.813 Found net devices under 0000:31:00.0: cvl_0_0 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:06:30.813 Found net devices under 0000:31:00.1: cvl_0_1 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # is_hw=yes 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:06:30.813 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:06:31.074 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:06:31.334 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:06:31.334 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.651 ms 00:06:31.334 00:06:31.335 --- 10.0.0.2 ping statistics --- 00:06:31.335 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:31.335 rtt min/avg/max/mdev = 0.651/0.651/0.651/0.000 ms 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:06:31.335 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:06:31.335 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.204 ms 00:06:31.335 00:06:31.335 --- 10.0.0.1 ping statistics --- 00:06:31.335 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:31.335 rtt min/avg/max/mdev = 0.204/0.204/0.204/0.000 ms 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@450 -- # return 0 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # nvmfpid=3640113 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # waitforlisten 3640113 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # '[' -z 3640113 ']' 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:31.335 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:31.335 19:02:43 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:06:31.335 [2024-11-26 19:02:43.818223] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:06:31.335 [2024-11-26 19:02:43.818288] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:31.335 [2024-11-26 19:02:43.907715] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:31.335 [2024-11-26 19:02:43.947754] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:06:31.335 [2024-11-26 19:02:43.947790] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:06:31.335 [2024-11-26 19:02:43.947798] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:31.335 [2024-11-26 19:02:43.947809] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:31.335 [2024-11-26 19:02:43.947815] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:06:31.335 [2024-11-26 19:02:43.949084] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:31.335 [2024-11-26 19:02:43.949172] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@868 -- # return 0 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:32.277 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:32.278 [2024-11-26 19:02:44.652258] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:32.278 [2024-11-26 19:02:44.668394] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:32.278 NULL1 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:32.278 Delay0 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=3640172 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:06:32.278 19:02:44 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:06:32.278 [2024-11-26 19:02:44.763239] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:06:34.190 19:02:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:06:34.190 19:02:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.190 19:02:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 starting I/O failed: -6 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 [2024-11-26 19:02:46.886964] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22c52c0 is same with the state(6) to be set 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.451 Write completed with error (sct=0, sc=8) 00:06:34.451 Read completed with error (sct=0, sc=8) 00:06:34.452 [2024-11-26 19:02:46.887361] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22c4f00 is same with the state(6) to be set 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 starting I/O failed: -6 00:06:34.452 [2024-11-26 19:02:46.892020] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f2ec800d4b0 is same with the state(6) to be set 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:34.452 Write completed with error (sct=0, sc=8) 00:06:34.452 Read completed with error (sct=0, sc=8) 00:06:35.395 [2024-11-26 19:02:47.862798] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22c65f0 is same with the state(6) to be set 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 [2024-11-26 19:02:47.890260] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22c50e0 is same with the state(6) to be set 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 [2024-11-26 19:02:47.890641] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22c54a0 is same with the state(6) to be set 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 [2024-11-26 19:02:47.893146] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f2ec800d7e0 is same with the state(6) to be set 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Write completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.395 Read completed with error (sct=0, sc=8) 00:06:35.396 Read completed with error (sct=0, sc=8) 00:06:35.396 Write completed with error (sct=0, sc=8) 00:06:35.396 Read completed with error (sct=0, sc=8) 00:06:35.396 Read completed with error (sct=0, sc=8) 00:06:35.396 Read completed with error (sct=0, sc=8) 00:06:35.396 Read completed with error (sct=0, sc=8) 00:06:35.396 Read completed with error (sct=0, sc=8) 00:06:35.396 Read completed with error (sct=0, sc=8) 00:06:35.396 Write completed with error (sct=0, sc=8) 00:06:35.396 [2024-11-26 19:02:47.894437] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f2ec800d020 is same with the state(6) to be set 00:06:35.396 Initializing NVMe Controllers 00:06:35.396 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:06:35.396 Controller IO queue size 128, less than required. 00:06:35.396 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:06:35.396 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:06:35.396 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:06:35.396 Initialization complete. Launching workers. 00:06:35.396 ======================================================== 00:06:35.396 Latency(us) 00:06:35.396 Device Information : IOPS MiB/s Average min max 00:06:35.396 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 163.43 0.08 908364.32 404.15 1005367.26 00:06:35.396 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 151.97 0.07 978379.75 264.24 2002516.24 00:06:35.396 ======================================================== 00:06:35.396 Total : 315.41 0.15 942100.04 264.24 2002516.24 00:06:35.396 00:06:35.396 [2024-11-26 19:02:47.895155] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22c65f0 (9): Bad file descriptor 00:06:35.396 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:06:35.396 19:02:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:35.396 19:02:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:06:35.396 19:02:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 3640172 00:06:35.396 19:02:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 3640172 00:06:35.965 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (3640172) - No such process 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 3640172 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # local es=0 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@654 -- # valid_exec_arg wait 3640172 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@640 -- # local arg=wait 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # type -t wait 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # wait 3640172 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # es=1 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:35.965 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:35.966 [2024-11-26 19:02:48.427226] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=3641029 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:35.966 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:06:35.966 [2024-11-26 19:02:48.513783] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:06:36.536 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:06:36.536 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:36.536 19:02:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:06:37.105 19:02:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:06:37.105 19:02:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:37.105 19:02:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:06:37.366 19:02:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:06:37.366 19:02:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:37.366 19:02:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:06:37.938 19:02:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:06:37.938 19:02:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:37.938 19:02:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:06:38.507 19:02:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:06:38.507 19:02:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:38.507 19:02:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:06:39.080 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:06:39.080 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:39.080 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:06:39.080 Initializing NVMe Controllers 00:06:39.080 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:06:39.080 Controller IO queue size 128, less than required. 00:06:39.080 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:06:39.080 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:06:39.080 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:06:39.080 Initialization complete. Launching workers. 00:06:39.080 ======================================================== 00:06:39.080 Latency(us) 00:06:39.080 Device Information : IOPS MiB/s Average min max 00:06:39.080 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1002375.52 1000236.80 1041312.00 00:06:39.080 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1003482.22 1000198.14 1041560.00 00:06:39.080 ======================================================== 00:06:39.080 Total : 256.00 0.12 1002928.87 1000198.14 1041560.00 00:06:39.080 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 3641029 00:06:39.654 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (3641029) - No such process 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 3641029 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # nvmfcleanup 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:06:39.654 19:02:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:06:39.654 rmmod nvme_tcp 00:06:39.654 rmmod nvme_fabrics 00:06:39.654 rmmod nvme_keyring 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@517 -- # '[' -n 3640113 ']' 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # killprocess 3640113 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # '[' -z 3640113 ']' 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@958 -- # kill -0 3640113 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # uname 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3640113 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3640113' 00:06:39.654 killing process with pid 3640113 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@973 -- # kill 3640113 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@978 -- # wait 3640113 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-save 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-restore 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:39.654 19:02:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:06:42.201 00:06:42.201 real 0m19.281s 00:06:42.201 user 0m30.908s 00:06:42.201 sys 0m7.547s 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:06:42.201 ************************************ 00:06:42.201 END TEST nvmf_delete_subsystem 00:06:42.201 ************************************ 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:06:42.201 ************************************ 00:06:42.201 START TEST nvmf_host_management 00:06:42.201 ************************************ 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:06:42.201 * Looking for test storage... 00:06:42.201 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1693 -- # lcov --version 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:42.201 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:42.201 --rc genhtml_branch_coverage=1 00:06:42.201 --rc genhtml_function_coverage=1 00:06:42.201 --rc genhtml_legend=1 00:06:42.201 --rc geninfo_all_blocks=1 00:06:42.201 --rc geninfo_unexecuted_blocks=1 00:06:42.201 00:06:42.201 ' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:42.201 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:42.201 --rc genhtml_branch_coverage=1 00:06:42.201 --rc genhtml_function_coverage=1 00:06:42.201 --rc genhtml_legend=1 00:06:42.201 --rc geninfo_all_blocks=1 00:06:42.201 --rc geninfo_unexecuted_blocks=1 00:06:42.201 00:06:42.201 ' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:42.201 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:42.201 --rc genhtml_branch_coverage=1 00:06:42.201 --rc genhtml_function_coverage=1 00:06:42.201 --rc genhtml_legend=1 00:06:42.201 --rc geninfo_all_blocks=1 00:06:42.201 --rc geninfo_unexecuted_blocks=1 00:06:42.201 00:06:42.201 ' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:42.201 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:42.201 --rc genhtml_branch_coverage=1 00:06:42.201 --rc genhtml_function_coverage=1 00:06:42.201 --rc genhtml_legend=1 00:06:42.201 --rc geninfo_all_blocks=1 00:06:42.201 --rc geninfo_unexecuted_blocks=1 00:06:42.201 00:06:42.201 ' 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:42.201 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:42.202 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@476 -- # prepare_net_devs 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@438 -- # local -g is_hw=no 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # remove_spdk_ns 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:06:42.202 19:02:54 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:06:50.439 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:06:50.440 Found 0000:31:00.0 (0x8086 - 0x159b) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:06:50.440 Found 0000:31:00.1 (0x8086 - 0x159b) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:06:50.440 Found net devices under 0000:31:00.0: cvl_0_0 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:06:50.440 Found net devices under 0000:31:00.1: cvl_0_1 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # is_hw=yes 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:06:50.440 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:06:50.440 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.691 ms 00:06:50.440 00:06:50.440 --- 10.0.0.2 ping statistics --- 00:06:50.440 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:50.440 rtt min/avg/max/mdev = 0.691/0.691/0.691/0.000 ms 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:06:50.440 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:06:50.440 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.319 ms 00:06:50.440 00:06:50.440 --- 10.0.0.1 ping statistics --- 00:06:50.440 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:50.440 rtt min/avg/max/mdev = 0.319/0.319/0.319/0.000 ms 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@450 -- # return 0 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:06:50.440 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@509 -- # nvmfpid=3646636 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@510 -- # waitforlisten 3646636 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 3646636 ']' 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:50.441 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:50.441 19:03:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:50.441 [2024-11-26 19:03:03.031834] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:06:50.441 [2024-11-26 19:03:03.031906] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:50.702 [2024-11-26 19:03:03.142077] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:50.702 [2024-11-26 19:03:03.194524] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:06:50.702 [2024-11-26 19:03:03.194580] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:06:50.702 [2024-11-26 19:03:03.194594] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:50.702 [2024-11-26 19:03:03.194602] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:50.702 [2024-11-26 19:03:03.194607] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:06:50.702 [2024-11-26 19:03:03.196639] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:50.702 [2024-11-26 19:03:03.196803] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:50.702 [2024-11-26 19:03:03.196957] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:06:50.702 [2024-11-26 19:03:03.197164] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.273 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:51.273 [2024-11-26 19:03:03.890381] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:51.534 Malloc0 00:06:51.534 [2024-11-26 19:03:03.964243] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:51.534 19:03:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=3646892 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 3646892 /var/tmp/bdevperf.sock 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 3646892 ']' 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:06:51.534 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:06:51.534 { 00:06:51.534 "params": { 00:06:51.534 "name": "Nvme$subsystem", 00:06:51.534 "trtype": "$TEST_TRANSPORT", 00:06:51.534 "traddr": "$NVMF_FIRST_TARGET_IP", 00:06:51.534 "adrfam": "ipv4", 00:06:51.534 "trsvcid": "$NVMF_PORT", 00:06:51.534 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:06:51.534 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:06:51.534 "hdgst": ${hdgst:-false}, 00:06:51.534 "ddgst": ${ddgst:-false} 00:06:51.534 }, 00:06:51.534 "method": "bdev_nvme_attach_controller" 00:06:51.534 } 00:06:51.534 EOF 00:06:51.534 )") 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:06:51.534 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:06:51.534 "params": { 00:06:51.534 "name": "Nvme0", 00:06:51.534 "trtype": "tcp", 00:06:51.534 "traddr": "10.0.0.2", 00:06:51.534 "adrfam": "ipv4", 00:06:51.534 "trsvcid": "4420", 00:06:51.534 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:06:51.534 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:06:51.534 "hdgst": false, 00:06:51.534 "ddgst": false 00:06:51.534 }, 00:06:51.534 "method": "bdev_nvme_attach_controller" 00:06:51.534 }' 00:06:51.534 [2024-11-26 19:03:04.077026] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:06:51.534 [2024-11-26 19:03:04.077085] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3646892 ] 00:06:51.534 [2024-11-26 19:03:04.155691] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:51.796 [2024-11-26 19:03:04.191895] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:52.056 Running I/O for 10 seconds... 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=579 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@58 -- # '[' 579 -ge 100 ']' 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@60 -- # break 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.317 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:52.579 [2024-11-26 19:03:04.939354] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.579 [2024-11-26 19:03:04.939402] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.579 [2024-11-26 19:03:04.939411] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.579 [2024-11-26 19:03:04.939418] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939426] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939433] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939440] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939447] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939454] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939461] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939468] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939475] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939482] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939489] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939501] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939508] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939515] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939522] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939529] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939536] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939543] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939550] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939557] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939563] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939570] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939577] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939583] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939590] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939597] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939604] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939610] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939617] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939623] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939630] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939637] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939644] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939651] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939658] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939665] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939671] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939678] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939687] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939693] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939700] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939707] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939714] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939721] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939727] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939734] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939740] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939747] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939753] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939760] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939767] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939774] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939781] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939788] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939795] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939802] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939809] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939816] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939823] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.939829] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x110a1b0 is same with the state(6) to be set 00:06:52.580 [2024-11-26 19:03:04.940421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:81920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:82048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:82176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:82304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:82432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:82560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:82688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:82816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:82944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.580 [2024-11-26 19:03:04.940625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:83072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.580 [2024-11-26 19:03:04.940633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:83200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:83328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:83456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:83584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:83712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:83840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:83968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:84096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:84224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:84352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:84480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:84608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940842] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:84736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:84864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:84992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:85120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:85248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:85376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:85504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:85632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.940988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.940997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:85760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:85888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:86016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:86144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:86272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:86400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:86528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:86656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:86784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:86912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:87040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:87168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:87296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:87424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.581 [2024-11-26 19:03:04.941241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:87552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.581 [2024-11-26 19:03:04.941250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:87680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:87808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:87936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:88064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:88192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:88320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:88448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:88576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:88704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:88832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:88960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:89088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:89216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:89344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:89472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:89600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:89728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:89856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:89984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:52.582 [2024-11-26 19:03:04.941580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.941589] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x135a270 is same with the state(6) to be set 00:06:52.582 [2024-11-26 19:03:04.942915] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:06:52.582 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.582 task offset: 81920 on job bdev=Nvme0n1 fails 00:06:52.582 00:06:52.582 Latency(us) 00:06:52.582 [2024-11-26T18:03:05.207Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:06:52.582 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:06:52.582 Job: Nvme0n1 ended in about 0.44 seconds with error 00:06:52.582 Verification LBA range: start 0x0 length 0x400 00:06:52.582 Nvme0n1 : 0.44 1457.49 91.09 145.75 0.00 38774.52 5079.04 34515.63 00:06:52.582 [2024-11-26T18:03:05.207Z] =================================================================================================================== 00:06:52.582 [2024-11-26T18:03:05.207Z] Total : 1457.49 91.09 145.75 0.00 38774.52 5079.04 34515.63 00:06:52.582 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:06:52.582 [2024-11-26 19:03:04.944950] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:52.582 [2024-11-26 19:03:04.944975] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1349b10 (9): Bad file descriptor 00:06:52.582 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.582 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:52.582 [2024-11-26 19:03:04.947165] ctrlr.c: 825:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode0' does not allow host 'nqn.2016-06.io.spdk:host0' 00:06:52.582 [2024-11-26 19:03:04.947289] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:3 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:06:52.582 [2024-11-26 19:03:04.947310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND SPECIFIC (01/84) qid:0 cid:3 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:52.582 [2024-11-26 19:03:04.947327] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode0 00:06:52.582 [2024-11-26 19:03:04.947335] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 132 00:06:52.582 [2024-11-26 19:03:04.947343] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:06:52.582 [2024-11-26 19:03:04.947350] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1349b10 00:06:52.582 [2024-11-26 19:03:04.947369] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1349b10 (9): Bad file descriptor 00:06:52.582 [2024-11-26 19:03:04.947383] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:06:52.582 [2024-11-26 19:03:04.947391] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:06:52.582 [2024-11-26 19:03:04.947400] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:06:52.582 [2024-11-26 19:03:04.947410] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:06:52.582 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.582 19:03:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 3646892 00:06:53.527 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (3646892) - No such process 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # true 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:06:53.527 { 00:06:53.527 "params": { 00:06:53.527 "name": "Nvme$subsystem", 00:06:53.527 "trtype": "$TEST_TRANSPORT", 00:06:53.527 "traddr": "$NVMF_FIRST_TARGET_IP", 00:06:53.527 "adrfam": "ipv4", 00:06:53.527 "trsvcid": "$NVMF_PORT", 00:06:53.527 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:06:53.527 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:06:53.527 "hdgst": ${hdgst:-false}, 00:06:53.527 "ddgst": ${ddgst:-false} 00:06:53.527 }, 00:06:53.527 "method": "bdev_nvme_attach_controller" 00:06:53.527 } 00:06:53.527 EOF 00:06:53.527 )") 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:06:53.527 19:03:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:06:53.527 "params": { 00:06:53.527 "name": "Nvme0", 00:06:53.527 "trtype": "tcp", 00:06:53.527 "traddr": "10.0.0.2", 00:06:53.527 "adrfam": "ipv4", 00:06:53.527 "trsvcid": "4420", 00:06:53.527 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:06:53.527 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:06:53.527 "hdgst": false, 00:06:53.527 "ddgst": false 00:06:53.527 }, 00:06:53.527 "method": "bdev_nvme_attach_controller" 00:06:53.527 }' 00:06:53.527 [2024-11-26 19:03:06.025113] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:06:53.527 [2024-11-26 19:03:06.025179] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3647358 ] 00:06:53.527 [2024-11-26 19:03:06.105471] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:53.527 [2024-11-26 19:03:06.141497] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:53.788 Running I/O for 1 seconds... 00:06:54.995 1854.00 IOPS, 115.88 MiB/s 00:06:54.995 Latency(us) 00:06:54.995 [2024-11-26T18:03:07.620Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:06:54.995 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:06:54.995 Verification LBA range: start 0x0 length 0x400 00:06:54.995 Nvme0n1 : 1.03 1860.03 116.25 0.00 0.00 33752.96 5434.03 30365.01 00:06:54.995 [2024-11-26T18:03:07.620Z] =================================================================================================================== 00:06:54.995 [2024-11-26T18:03:07.620Z] Total : 1860.03 116.25 0.00 0.00 33752.96 5434.03 30365.01 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@516 -- # nvmfcleanup 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:06:54.995 rmmod nvme_tcp 00:06:54.995 rmmod nvme_fabrics 00:06:54.995 rmmod nvme_keyring 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@517 -- # '[' -n 3646636 ']' 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@518 -- # killprocess 3646636 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@954 -- # '[' -z 3646636 ']' 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@958 -- # kill -0 3646636 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@959 -- # uname 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:54.995 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3646636 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3646636' 00:06:55.256 killing process with pid 3646636 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@973 -- # kill 3646636 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@978 -- # wait 3646636 00:06:55.256 [2024-11-26 19:03:07.732347] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-save 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-restore 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:55.256 19:03:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:06:57.818 00:06:57.818 real 0m15.431s 00:06:57.818 user 0m23.206s 00:06:57.818 sys 0m7.191s 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:57.818 ************************************ 00:06:57.818 END TEST nvmf_host_management 00:06:57.818 ************************************ 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:06:57.818 ************************************ 00:06:57.818 START TEST nvmf_lvol 00:06:57.818 ************************************ 00:06:57.818 19:03:09 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:06:57.818 * Looking for test storage... 00:06:57.818 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1693 -- # lcov --version 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:57.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.818 --rc genhtml_branch_coverage=1 00:06:57.818 --rc genhtml_function_coverage=1 00:06:57.818 --rc genhtml_legend=1 00:06:57.818 --rc geninfo_all_blocks=1 00:06:57.818 --rc geninfo_unexecuted_blocks=1 00:06:57.818 00:06:57.818 ' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:57.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.818 --rc genhtml_branch_coverage=1 00:06:57.818 --rc genhtml_function_coverage=1 00:06:57.818 --rc genhtml_legend=1 00:06:57.818 --rc geninfo_all_blocks=1 00:06:57.818 --rc geninfo_unexecuted_blocks=1 00:06:57.818 00:06:57.818 ' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:57.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.818 --rc genhtml_branch_coverage=1 00:06:57.818 --rc genhtml_function_coverage=1 00:06:57.818 --rc genhtml_legend=1 00:06:57.818 --rc geninfo_all_blocks=1 00:06:57.818 --rc geninfo_unexecuted_blocks=1 00:06:57.818 00:06:57.818 ' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:57.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.818 --rc genhtml_branch_coverage=1 00:06:57.818 --rc genhtml_function_coverage=1 00:06:57.818 --rc genhtml_legend=1 00:06:57.818 --rc geninfo_all_blocks=1 00:06:57.818 --rc geninfo_unexecuted_blocks=1 00:06:57.818 00:06:57.818 ' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:57.818 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:57.819 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@476 -- # prepare_net_devs 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@438 -- # local -g is_hw=no 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # remove_spdk_ns 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:06:57.819 19:03:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:07:05.968 Found 0000:31:00.0 (0x8086 - 0x159b) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:07:05.968 Found 0000:31:00.1 (0x8086 - 0x159b) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:07:05.968 Found net devices under 0000:31:00.0: cvl_0_0 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:07:05.968 Found net devices under 0000:31:00.1: cvl_0_1 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # is_hw=yes 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:07:05.968 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:05.969 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:05.969 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:05.969 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:07:05.969 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:07:05.969 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:05.969 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.351 ms 00:07:05.969 00:07:05.969 --- 10.0.0.2 ping statistics --- 00:07:05.969 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:05.969 rtt min/avg/max/mdev = 0.351/0.351/0.351/0.000 ms 00:07:05.969 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:06.232 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:06.232 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.316 ms 00:07:06.232 00:07:06.232 --- 10.0.0.1 ping statistics --- 00:07:06.232 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:06.232 rtt min/avg/max/mdev = 0.316/0.316/0.316/0.000 ms 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@450 -- # return 0 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@726 -- # xtrace_disable 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@509 -- # nvmfpid=3652857 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@510 -- # waitforlisten 3652857 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@835 -- # '[' -z 3652857 ']' 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:06.232 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:06.232 19:03:18 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:07:06.232 [2024-11-26 19:03:18.699095] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:07:06.232 [2024-11-26 19:03:18.699162] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:06.232 [2024-11-26 19:03:18.789127] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:06.232 [2024-11-26 19:03:18.829898] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:06.232 [2024-11-26 19:03:18.829935] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:06.232 [2024-11-26 19:03:18.829943] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:06.232 [2024-11-26 19:03:18.829950] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:06.232 [2024-11-26 19:03:18.829955] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:06.232 [2024-11-26 19:03:18.831569] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:06.232 [2024-11-26 19:03:18.831685] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:06.232 [2024-11-26 19:03:18.831687] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@868 -- # return 0 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@732 -- # xtrace_disable 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:07:07.173 [2024-11-26 19:03:19.695477] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:07.173 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:07:07.435 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:07:07.435 19:03:19 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:07:07.695 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:07:07.695 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:07:07.695 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:07:07.956 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=a47785f8-bb9e-4fec-80a3-e38e548936c8 00:07:07.956 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u a47785f8-bb9e-4fec-80a3-e38e548936c8 lvol 20 00:07:08.217 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=86bc47a6-9dd8-4737-8359-7c5f33815204 00:07:08.217 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:07:08.478 19:03:20 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 86bc47a6-9dd8-4737-8359-7c5f33815204 00:07:08.478 19:03:21 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:07:08.738 [2024-11-26 19:03:21.224531] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:08.738 19:03:21 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:09.000 19:03:21 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=3653517 00:07:09.000 19:03:21 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:07:09.000 19:03:21 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:07:09.945 19:03:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot 86bc47a6-9dd8-4737-8359-7c5f33815204 MY_SNAPSHOT 00:07:10.207 19:03:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=4c6adacc-cf7c-4304-9625-eff7c1942463 00:07:10.207 19:03:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize 86bc47a6-9dd8-4737-8359-7c5f33815204 30 00:07:10.469 19:03:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone 4c6adacc-cf7c-4304-9625-eff7c1942463 MY_CLONE 00:07:10.730 19:03:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=7c393c75-4c2d-4379-97e9-0e364f6ac18a 00:07:10.730 19:03:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate 7c393c75-4c2d-4379-97e9-0e364f6ac18a 00:07:11.303 19:03:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 3653517 00:07:19.448 Initializing NVMe Controllers 00:07:19.448 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:07:19.448 Controller IO queue size 128, less than required. 00:07:19.448 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:07:19.448 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:07:19.448 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:07:19.448 Initialization complete. Launching workers. 00:07:19.448 ======================================================== 00:07:19.448 Latency(us) 00:07:19.448 Device Information : IOPS MiB/s Average min max 00:07:19.448 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 12105.60 47.29 10576.24 1491.85 59148.45 00:07:19.448 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 17698.90 69.14 7232.79 347.15 54165.47 00:07:19.448 ======================================================== 00:07:19.448 Total : 29804.50 116.42 8590.79 347.15 59148.45 00:07:19.448 00:07:19.448 19:03:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:07:19.448 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 86bc47a6-9dd8-4737-8359-7c5f33815204 00:07:19.709 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u a47785f8-bb9e-4fec-80a3-e38e548936c8 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@516 -- # nvmfcleanup 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:07:19.970 rmmod nvme_tcp 00:07:19.970 rmmod nvme_fabrics 00:07:19.970 rmmod nvme_keyring 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@517 -- # '[' -n 3652857 ']' 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@518 -- # killprocess 3652857 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@954 -- # '[' -z 3652857 ']' 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@958 -- # kill -0 3652857 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@959 -- # uname 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3652857 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3652857' 00:07:19.970 killing process with pid 3652857 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@973 -- # kill 3652857 00:07:19.970 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@978 -- # wait 3652857 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-save 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-restore 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:20.232 19:03:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:22.146 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:07:22.146 00:07:22.146 real 0m24.860s 00:07:22.146 user 1m4.759s 00:07:22.146 sys 0m9.340s 00:07:22.146 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:22.146 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:07:22.146 ************************************ 00:07:22.146 END TEST nvmf_lvol 00:07:22.146 ************************************ 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:07:22.408 ************************************ 00:07:22.408 START TEST nvmf_lvs_grow 00:07:22.408 ************************************ 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:07:22.408 * Looking for test storage... 00:07:22.408 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1693 -- # lcov --version 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:22.408 19:03:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:22.408 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.408 --rc genhtml_branch_coverage=1 00:07:22.408 --rc genhtml_function_coverage=1 00:07:22.408 --rc genhtml_legend=1 00:07:22.408 --rc geninfo_all_blocks=1 00:07:22.408 --rc geninfo_unexecuted_blocks=1 00:07:22.408 00:07:22.408 ' 00:07:22.408 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:22.408 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.408 --rc genhtml_branch_coverage=1 00:07:22.408 --rc genhtml_function_coverage=1 00:07:22.408 --rc genhtml_legend=1 00:07:22.408 --rc geninfo_all_blocks=1 00:07:22.408 --rc geninfo_unexecuted_blocks=1 00:07:22.408 00:07:22.408 ' 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:22.409 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.409 --rc genhtml_branch_coverage=1 00:07:22.409 --rc genhtml_function_coverage=1 00:07:22.409 --rc genhtml_legend=1 00:07:22.409 --rc geninfo_all_blocks=1 00:07:22.409 --rc geninfo_unexecuted_blocks=1 00:07:22.409 00:07:22.409 ' 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:22.409 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.409 --rc genhtml_branch_coverage=1 00:07:22.409 --rc genhtml_function_coverage=1 00:07:22.409 --rc genhtml_legend=1 00:07:22.409 --rc geninfo_all_blocks=1 00:07:22.409 --rc geninfo_unexecuted_blocks=1 00:07:22.409 00:07:22.409 ' 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:22.409 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:22.670 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:07:22.670 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:07:22.670 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:22.670 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:22.670 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:22.670 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:07:22.671 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@476 -- # prepare_net_devs 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@438 -- # local -g is_hw=no 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # remove_spdk_ns 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:07:22.671 19:03:35 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:07:30.821 Found 0000:31:00.0 (0x8086 - 0x159b) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:07:30.821 Found 0000:31:00.1 (0x8086 - 0x159b) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:07:30.821 Found net devices under 0000:31:00.0: cvl_0_0 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:07:30.821 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:07:30.822 Found net devices under 0000:31:00.1: cvl_0_1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # is_hw=yes 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:07:30.822 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:30.822 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.620 ms 00:07:30.822 00:07:30.822 --- 10.0.0.2 ping statistics --- 00:07:30.822 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:30.822 rtt min/avg/max/mdev = 0.620/0.620/0.620/0.000 ms 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:30.822 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:30.822 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.310 ms 00:07:30.822 00:07:30.822 --- 10.0.0.1 ping statistics --- 00:07:30.822 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:30.822 rtt min/avg/max/mdev = 0.310/0.310/0.310/0.000 ms 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@450 -- # return 0 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@726 -- # xtrace_disable 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@509 -- # nvmfpid=3660377 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@510 -- # waitforlisten 3660377 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # '[' -z 3660377 ']' 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:30.822 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:30.822 19:03:43 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:31.084 [2024-11-26 19:03:43.470080] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:07:31.084 [2024-11-26 19:03:43.470131] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:31.084 [2024-11-26 19:03:43.556029] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:31.084 [2024-11-26 19:03:43.590937] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:31.084 [2024-11-26 19:03:43.590971] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:31.084 [2024-11-26 19:03:43.590979] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:31.084 [2024-11-26 19:03:43.590985] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:31.084 [2024-11-26 19:03:43.590991] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:31.084 [2024-11-26 19:03:43.591572] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:31.657 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:31.657 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@868 -- # return 0 00:07:31.657 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:07:31.657 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@732 -- # xtrace_disable 00:07:31.657 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:07:31.918 [2024-11-26 19:03:44.461308] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:31.918 ************************************ 00:07:31.918 START TEST lvs_grow_clean 00:07:31.918 ************************************ 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1129 -- # lvs_grow 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:31.918 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:32.178 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:07:32.178 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:07:32.178 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:07:32.439 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=70417bfc-2b89-43f2-b037-502818f9f482 00:07:32.439 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:32.439 19:03:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:07:32.700 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:07:32.700 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:07:32.700 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 70417bfc-2b89-43f2-b037-502818f9f482 lvol 150 00:07:32.700 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=35dc34a3-2a64-445a-8d1a-40a197ee3b5a 00:07:32.700 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:32.700 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:07:32.960 [2024-11-26 19:03:45.408526] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:07:32.960 [2024-11-26 19:03:45.408579] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:07:32.960 true 00:07:32.960 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:07:32.960 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:33.220 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:07:33.220 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:07:33.220 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 35dc34a3-2a64-445a-8d1a-40a197ee3b5a 00:07:33.481 19:03:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:07:33.481 [2024-11-26 19:03:46.054540] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:33.481 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:33.742 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=3660997 00:07:33.742 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 3660997 /var/tmp/bdevperf.sock 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # '[' -z 3660997 ']' 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:07:33.743 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:33.743 19:03:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:07:33.743 [2024-11-26 19:03:46.271295] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:07:33.743 [2024-11-26 19:03:46.271345] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3660997 ] 00:07:34.004 [2024-11-26 19:03:46.366678] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:34.004 [2024-11-26 19:03:46.402518] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:34.576 19:03:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:34.576 19:03:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@868 -- # return 0 00:07:34.576 19:03:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:07:34.837 Nvme0n1 00:07:34.837 19:03:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:07:35.099 [ 00:07:35.099 { 00:07:35.099 "name": "Nvme0n1", 00:07:35.099 "aliases": [ 00:07:35.099 "35dc34a3-2a64-445a-8d1a-40a197ee3b5a" 00:07:35.099 ], 00:07:35.099 "product_name": "NVMe disk", 00:07:35.099 "block_size": 4096, 00:07:35.099 "num_blocks": 38912, 00:07:35.099 "uuid": "35dc34a3-2a64-445a-8d1a-40a197ee3b5a", 00:07:35.099 "numa_id": 0, 00:07:35.099 "assigned_rate_limits": { 00:07:35.099 "rw_ios_per_sec": 0, 00:07:35.099 "rw_mbytes_per_sec": 0, 00:07:35.099 "r_mbytes_per_sec": 0, 00:07:35.099 "w_mbytes_per_sec": 0 00:07:35.099 }, 00:07:35.099 "claimed": false, 00:07:35.099 "zoned": false, 00:07:35.099 "supported_io_types": { 00:07:35.099 "read": true, 00:07:35.099 "write": true, 00:07:35.099 "unmap": true, 00:07:35.099 "flush": true, 00:07:35.099 "reset": true, 00:07:35.099 "nvme_admin": true, 00:07:35.099 "nvme_io": true, 00:07:35.099 "nvme_io_md": false, 00:07:35.099 "write_zeroes": true, 00:07:35.099 "zcopy": false, 00:07:35.099 "get_zone_info": false, 00:07:35.099 "zone_management": false, 00:07:35.099 "zone_append": false, 00:07:35.099 "compare": true, 00:07:35.099 "compare_and_write": true, 00:07:35.099 "abort": true, 00:07:35.099 "seek_hole": false, 00:07:35.099 "seek_data": false, 00:07:35.099 "copy": true, 00:07:35.099 "nvme_iov_md": false 00:07:35.099 }, 00:07:35.099 "memory_domains": [ 00:07:35.099 { 00:07:35.099 "dma_device_id": "system", 00:07:35.099 "dma_device_type": 1 00:07:35.099 } 00:07:35.099 ], 00:07:35.099 "driver_specific": { 00:07:35.099 "nvme": [ 00:07:35.099 { 00:07:35.099 "trid": { 00:07:35.099 "trtype": "TCP", 00:07:35.099 "adrfam": "IPv4", 00:07:35.099 "traddr": "10.0.0.2", 00:07:35.099 "trsvcid": "4420", 00:07:35.099 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:07:35.099 }, 00:07:35.099 "ctrlr_data": { 00:07:35.099 "cntlid": 1, 00:07:35.099 "vendor_id": "0x8086", 00:07:35.099 "model_number": "SPDK bdev Controller", 00:07:35.099 "serial_number": "SPDK0", 00:07:35.099 "firmware_revision": "25.01", 00:07:35.099 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:07:35.099 "oacs": { 00:07:35.099 "security": 0, 00:07:35.099 "format": 0, 00:07:35.099 "firmware": 0, 00:07:35.099 "ns_manage": 0 00:07:35.099 }, 00:07:35.099 "multi_ctrlr": true, 00:07:35.099 "ana_reporting": false 00:07:35.099 }, 00:07:35.099 "vs": { 00:07:35.099 "nvme_version": "1.3" 00:07:35.099 }, 00:07:35.099 "ns_data": { 00:07:35.099 "id": 1, 00:07:35.099 "can_share": true 00:07:35.099 } 00:07:35.099 } 00:07:35.099 ], 00:07:35.099 "mp_policy": "active_passive" 00:07:35.099 } 00:07:35.099 } 00:07:35.099 ] 00:07:35.099 19:03:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=3661340 00:07:35.100 19:03:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:07:35.100 19:03:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:07:35.100 Running I/O for 10 seconds... 00:07:36.040 Latency(us) 00:07:36.040 [2024-11-26T18:03:48.666Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:36.041 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:36.041 Nvme0n1 : 1.00 17603.00 68.76 0.00 0.00 0.00 0.00 0.00 00:07:36.041 [2024-11-26T18:03:48.666Z] =================================================================================================================== 00:07:36.041 [2024-11-26T18:03:48.666Z] Total : 17603.00 68.76 0.00 0.00 0.00 0.00 0.00 00:07:36.041 00:07:36.981 19:03:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:37.241 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:37.241 Nvme0n1 : 2.00 17753.50 69.35 0.00 0.00 0.00 0.00 0.00 00:07:37.241 [2024-11-26T18:03:49.866Z] =================================================================================================================== 00:07:37.241 [2024-11-26T18:03:49.866Z] Total : 17753.50 69.35 0.00 0.00 0.00 0.00 0.00 00:07:37.241 00:07:37.241 true 00:07:37.241 19:03:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:37.241 19:03:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:07:37.501 19:03:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:07:37.501 19:03:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:07:37.501 19:03:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 3661340 00:07:38.073 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:38.073 Nvme0n1 : 3.00 17693.67 69.12 0.00 0.00 0.00 0.00 0.00 00:07:38.073 [2024-11-26T18:03:50.698Z] =================================================================================================================== 00:07:38.073 [2024-11-26T18:03:50.698Z] Total : 17693.67 69.12 0.00 0.00 0.00 0.00 0.00 00:07:38.073 00:07:39.017 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:39.017 Nvme0n1 : 4.00 17746.25 69.32 0.00 0.00 0.00 0.00 0.00 00:07:39.017 [2024-11-26T18:03:51.642Z] =================================================================================================================== 00:07:39.017 [2024-11-26T18:03:51.642Z] Total : 17746.25 69.32 0.00 0.00 0.00 0.00 0.00 00:07:39.017 00:07:40.404 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:40.404 Nvme0n1 : 5.00 17805.80 69.55 0.00 0.00 0.00 0.00 0.00 00:07:40.404 [2024-11-26T18:03:53.029Z] =================================================================================================================== 00:07:40.404 [2024-11-26T18:03:53.029Z] Total : 17805.80 69.55 0.00 0.00 0.00 0.00 0.00 00:07:40.404 00:07:41.344 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:41.344 Nvme0n1 : 6.00 17841.67 69.69 0.00 0.00 0.00 0.00 0.00 00:07:41.344 [2024-11-26T18:03:53.969Z] =================================================================================================================== 00:07:41.344 [2024-11-26T18:03:53.969Z] Total : 17841.67 69.69 0.00 0.00 0.00 0.00 0.00 00:07:41.344 00:07:42.284 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:42.284 Nvme0n1 : 7.00 17869.86 69.80 0.00 0.00 0.00 0.00 0.00 00:07:42.284 [2024-11-26T18:03:54.909Z] =================================================================================================================== 00:07:42.284 [2024-11-26T18:03:54.909Z] Total : 17869.86 69.80 0.00 0.00 0.00 0.00 0.00 00:07:42.284 00:07:43.229 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:43.229 Nvme0n1 : 8.00 17894.25 69.90 0.00 0.00 0.00 0.00 0.00 00:07:43.229 [2024-11-26T18:03:55.854Z] =================================================================================================================== 00:07:43.229 [2024-11-26T18:03:55.854Z] Total : 17894.25 69.90 0.00 0.00 0.00 0.00 0.00 00:07:43.229 00:07:44.167 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:44.167 Nvme0n1 : 9.00 17904.89 69.94 0.00 0.00 0.00 0.00 0.00 00:07:44.167 [2024-11-26T18:03:56.792Z] =================================================================================================================== 00:07:44.167 [2024-11-26T18:03:56.792Z] Total : 17904.89 69.94 0.00 0.00 0.00 0.00 0.00 00:07:44.167 00:07:45.110 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:45.110 Nvme0n1 : 10.00 17923.90 70.02 0.00 0.00 0.00 0.00 0.00 00:07:45.110 [2024-11-26T18:03:57.735Z] =================================================================================================================== 00:07:45.110 [2024-11-26T18:03:57.735Z] Total : 17923.90 70.02 0.00 0.00 0.00 0.00 0.00 00:07:45.110 00:07:45.110 00:07:45.110 Latency(us) 00:07:45.110 [2024-11-26T18:03:57.735Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:45.110 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:45.110 Nvme0n1 : 10.01 17924.81 70.02 0.00 0.00 7136.71 3003.73 16711.68 00:07:45.110 [2024-11-26T18:03:57.735Z] =================================================================================================================== 00:07:45.110 [2024-11-26T18:03:57.735Z] Total : 17924.81 70.02 0.00 0.00 7136.71 3003.73 16711.68 00:07:45.110 { 00:07:45.110 "results": [ 00:07:45.110 { 00:07:45.110 "job": "Nvme0n1", 00:07:45.110 "core_mask": "0x2", 00:07:45.110 "workload": "randwrite", 00:07:45.110 "status": "finished", 00:07:45.110 "queue_depth": 128, 00:07:45.110 "io_size": 4096, 00:07:45.111 "runtime": 10.006635, 00:07:45.111 "iops": 17924.806890628068, 00:07:45.111 "mibps": 70.01877691651589, 00:07:45.111 "io_failed": 0, 00:07:45.111 "io_timeout": 0, 00:07:45.111 "avg_latency_us": 7136.711368014555, 00:07:45.111 "min_latency_us": 3003.733333333333, 00:07:45.111 "max_latency_us": 16711.68 00:07:45.111 } 00:07:45.111 ], 00:07:45.111 "core_count": 1 00:07:45.111 } 00:07:45.111 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 3660997 00:07:45.111 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # '[' -z 3660997 ']' 00:07:45.111 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@958 -- # kill -0 3660997 00:07:45.111 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # uname 00:07:45.111 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:45.111 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3660997 00:07:45.371 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:07:45.371 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:07:45.371 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3660997' 00:07:45.371 killing process with pid 3660997 00:07:45.371 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@973 -- # kill 3660997 00:07:45.371 Received shutdown signal, test time was about 10.000000 seconds 00:07:45.371 00:07:45.371 Latency(us) 00:07:45.371 [2024-11-26T18:03:57.996Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:45.371 [2024-11-26T18:03:57.996Z] =================================================================================================================== 00:07:45.371 [2024-11-26T18:03:57.996Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:07:45.371 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@978 -- # wait 3660997 00:07:45.371 19:03:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:45.631 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:07:45.631 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:45.631 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:07:45.891 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:07:45.891 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:07:45.891 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:07:45.891 [2024-11-26 19:03:58.499059] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # local es=0 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:46.157 request: 00:07:46.157 { 00:07:46.157 "uuid": "70417bfc-2b89-43f2-b037-502818f9f482", 00:07:46.157 "method": "bdev_lvol_get_lvstores", 00:07:46.157 "req_id": 1 00:07:46.157 } 00:07:46.157 Got JSON-RPC error response 00:07:46.157 response: 00:07:46.157 { 00:07:46.157 "code": -19, 00:07:46.157 "message": "No such device" 00:07:46.157 } 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # es=1 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:46.157 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:07:46.435 aio_bdev 00:07:46.435 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 35dc34a3-2a64-445a-8d1a-40a197ee3b5a 00:07:46.435 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@903 -- # local bdev_name=35dc34a3-2a64-445a-8d1a-40a197ee3b5a 00:07:46.435 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:46.435 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@905 -- # local i 00:07:46.436 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:46.436 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:46.436 19:03:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:07:46.805 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 35dc34a3-2a64-445a-8d1a-40a197ee3b5a -t 2000 00:07:46.805 [ 00:07:46.805 { 00:07:46.805 "name": "35dc34a3-2a64-445a-8d1a-40a197ee3b5a", 00:07:46.805 "aliases": [ 00:07:46.805 "lvs/lvol" 00:07:46.805 ], 00:07:46.805 "product_name": "Logical Volume", 00:07:46.805 "block_size": 4096, 00:07:46.805 "num_blocks": 38912, 00:07:46.805 "uuid": "35dc34a3-2a64-445a-8d1a-40a197ee3b5a", 00:07:46.805 "assigned_rate_limits": { 00:07:46.805 "rw_ios_per_sec": 0, 00:07:46.805 "rw_mbytes_per_sec": 0, 00:07:46.805 "r_mbytes_per_sec": 0, 00:07:46.805 "w_mbytes_per_sec": 0 00:07:46.805 }, 00:07:46.805 "claimed": false, 00:07:46.805 "zoned": false, 00:07:46.805 "supported_io_types": { 00:07:46.805 "read": true, 00:07:46.805 "write": true, 00:07:46.805 "unmap": true, 00:07:46.805 "flush": false, 00:07:46.805 "reset": true, 00:07:46.805 "nvme_admin": false, 00:07:46.805 "nvme_io": false, 00:07:46.805 "nvme_io_md": false, 00:07:46.805 "write_zeroes": true, 00:07:46.805 "zcopy": false, 00:07:46.805 "get_zone_info": false, 00:07:46.805 "zone_management": false, 00:07:46.805 "zone_append": false, 00:07:46.805 "compare": false, 00:07:46.805 "compare_and_write": false, 00:07:46.805 "abort": false, 00:07:46.805 "seek_hole": true, 00:07:46.805 "seek_data": true, 00:07:46.805 "copy": false, 00:07:46.805 "nvme_iov_md": false 00:07:46.805 }, 00:07:46.805 "driver_specific": { 00:07:46.805 "lvol": { 00:07:46.805 "lvol_store_uuid": "70417bfc-2b89-43f2-b037-502818f9f482", 00:07:46.805 "base_bdev": "aio_bdev", 00:07:46.805 "thin_provision": false, 00:07:46.805 "num_allocated_clusters": 38, 00:07:46.805 "snapshot": false, 00:07:46.805 "clone": false, 00:07:46.805 "esnap_clone": false 00:07:46.805 } 00:07:46.805 } 00:07:46.805 } 00:07:46.805 ] 00:07:46.805 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@911 -- # return 0 00:07:46.805 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:46.805 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:07:47.075 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:07:47.075 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:47.075 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:07:47.075 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:07:47.075 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 35dc34a3-2a64-445a-8d1a-40a197ee3b5a 00:07:47.336 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 70417bfc-2b89-43f2-b037-502818f9f482 00:07:47.336 19:03:59 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:47.597 00:07:47.597 real 0m15.588s 00:07:47.597 user 0m15.376s 00:07:47.597 sys 0m1.318s 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:07:47.597 ************************************ 00:07:47.597 END TEST lvs_grow_clean 00:07:47.597 ************************************ 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:47.597 ************************************ 00:07:47.597 START TEST lvs_grow_dirty 00:07:47.597 ************************************ 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1129 -- # lvs_grow dirty 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:47.597 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:07:47.857 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:07:47.857 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:07:48.118 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:07:48.118 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:07:48.118 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:07:48.382 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:07:48.382 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:07:48.382 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 lvol 150 00:07:48.382 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=c4d57573-9c7f-49d0-a4e0-c1a1821acf7f 00:07:48.382 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:48.382 19:04:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:07:48.644 [2024-11-26 19:04:01.091136] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:07:48.644 [2024-11-26 19:04:01.091192] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:07:48.644 true 00:07:48.644 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:07:48.644 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:07:48.904 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:07:48.904 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:07:48.904 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 c4d57573-9c7f-49d0-a4e0-c1a1821acf7f 00:07:49.165 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:07:49.165 [2024-11-26 19:04:01.737124] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:49.165 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=3664115 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 3664115 /var/tmp/bdevperf.sock 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 3664115 ']' 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:07:49.426 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:49.426 19:04:01 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:07:49.426 [2024-11-26 19:04:01.935381] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:07:49.426 [2024-11-26 19:04:01.935430] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3664115 ] 00:07:49.426 [2024-11-26 19:04:02.030945] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:49.690 [2024-11-26 19:04:02.066761] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:50.264 19:04:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:50.264 19:04:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:07:50.264 19:04:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:07:50.524 Nvme0n1 00:07:50.793 19:04:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:07:50.793 [ 00:07:50.793 { 00:07:50.793 "name": "Nvme0n1", 00:07:50.793 "aliases": [ 00:07:50.793 "c4d57573-9c7f-49d0-a4e0-c1a1821acf7f" 00:07:50.793 ], 00:07:50.793 "product_name": "NVMe disk", 00:07:50.793 "block_size": 4096, 00:07:50.793 "num_blocks": 38912, 00:07:50.793 "uuid": "c4d57573-9c7f-49d0-a4e0-c1a1821acf7f", 00:07:50.793 "numa_id": 0, 00:07:50.793 "assigned_rate_limits": { 00:07:50.793 "rw_ios_per_sec": 0, 00:07:50.793 "rw_mbytes_per_sec": 0, 00:07:50.793 "r_mbytes_per_sec": 0, 00:07:50.793 "w_mbytes_per_sec": 0 00:07:50.793 }, 00:07:50.793 "claimed": false, 00:07:50.793 "zoned": false, 00:07:50.793 "supported_io_types": { 00:07:50.793 "read": true, 00:07:50.793 "write": true, 00:07:50.793 "unmap": true, 00:07:50.793 "flush": true, 00:07:50.793 "reset": true, 00:07:50.793 "nvme_admin": true, 00:07:50.793 "nvme_io": true, 00:07:50.793 "nvme_io_md": false, 00:07:50.793 "write_zeroes": true, 00:07:50.793 "zcopy": false, 00:07:50.793 "get_zone_info": false, 00:07:50.793 "zone_management": false, 00:07:50.793 "zone_append": false, 00:07:50.793 "compare": true, 00:07:50.793 "compare_and_write": true, 00:07:50.793 "abort": true, 00:07:50.793 "seek_hole": false, 00:07:50.793 "seek_data": false, 00:07:50.793 "copy": true, 00:07:50.793 "nvme_iov_md": false 00:07:50.793 }, 00:07:50.793 "memory_domains": [ 00:07:50.793 { 00:07:50.793 "dma_device_id": "system", 00:07:50.793 "dma_device_type": 1 00:07:50.793 } 00:07:50.793 ], 00:07:50.793 "driver_specific": { 00:07:50.793 "nvme": [ 00:07:50.793 { 00:07:50.793 "trid": { 00:07:50.793 "trtype": "TCP", 00:07:50.793 "adrfam": "IPv4", 00:07:50.793 "traddr": "10.0.0.2", 00:07:50.793 "trsvcid": "4420", 00:07:50.793 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:07:50.793 }, 00:07:50.793 "ctrlr_data": { 00:07:50.793 "cntlid": 1, 00:07:50.793 "vendor_id": "0x8086", 00:07:50.793 "model_number": "SPDK bdev Controller", 00:07:50.793 "serial_number": "SPDK0", 00:07:50.793 "firmware_revision": "25.01", 00:07:50.794 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:07:50.794 "oacs": { 00:07:50.794 "security": 0, 00:07:50.794 "format": 0, 00:07:50.794 "firmware": 0, 00:07:50.794 "ns_manage": 0 00:07:50.794 }, 00:07:50.794 "multi_ctrlr": true, 00:07:50.794 "ana_reporting": false 00:07:50.794 }, 00:07:50.794 "vs": { 00:07:50.794 "nvme_version": "1.3" 00:07:50.794 }, 00:07:50.794 "ns_data": { 00:07:50.794 "id": 1, 00:07:50.794 "can_share": true 00:07:50.794 } 00:07:50.794 } 00:07:50.794 ], 00:07:50.794 "mp_policy": "active_passive" 00:07:50.794 } 00:07:50.794 } 00:07:50.794 ] 00:07:50.794 19:04:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=3664441 00:07:50.794 19:04:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:07:50.794 19:04:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:07:51.058 Running I/O for 10 seconds... 00:07:52.001 Latency(us) 00:07:52.001 [2024-11-26T18:04:04.626Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:52.001 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:52.001 Nvme0n1 : 1.00 17745.00 69.32 0.00 0.00 0.00 0.00 0.00 00:07:52.001 [2024-11-26T18:04:04.626Z] =================================================================================================================== 00:07:52.001 [2024-11-26T18:04:04.626Z] Total : 17745.00 69.32 0.00 0.00 0.00 0.00 0.00 00:07:52.001 00:07:52.946 19:04:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:07:52.946 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:52.946 Nvme0n1 : 2.00 17846.00 69.71 0.00 0.00 0.00 0.00 0.00 00:07:52.946 [2024-11-26T18:04:05.571Z] =================================================================================================================== 00:07:52.946 [2024-11-26T18:04:05.571Z] Total : 17846.00 69.71 0.00 0.00 0.00 0.00 0.00 00:07:52.946 00:07:52.946 true 00:07:52.946 19:04:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:07:52.946 19:04:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:07:53.207 19:04:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:07:53.207 19:04:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:07:53.207 19:04:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 3664441 00:07:54.151 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:54.151 Nvme0n1 : 3.00 17878.33 69.84 0.00 0.00 0.00 0.00 0.00 00:07:54.151 [2024-11-26T18:04:06.776Z] =================================================================================================================== 00:07:54.151 [2024-11-26T18:04:06.776Z] Total : 17878.33 69.84 0.00 0.00 0.00 0.00 0.00 00:07:54.151 00:07:55.094 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:55.094 Nvme0n1 : 4.00 17900.25 69.92 0.00 0.00 0.00 0.00 0.00 00:07:55.094 [2024-11-26T18:04:07.719Z] =================================================================================================================== 00:07:55.094 [2024-11-26T18:04:07.719Z] Total : 17900.25 69.92 0.00 0.00 0.00 0.00 0.00 00:07:55.094 00:07:56.036 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:56.036 Nvme0n1 : 5.00 17936.60 70.06 0.00 0.00 0.00 0.00 0.00 00:07:56.036 [2024-11-26T18:04:08.661Z] =================================================================================================================== 00:07:56.036 [2024-11-26T18:04:08.661Z] Total : 17936.60 70.06 0.00 0.00 0.00 0.00 0.00 00:07:56.036 00:07:56.979 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:56.979 Nvme0n1 : 6.00 17939.50 70.08 0.00 0.00 0.00 0.00 0.00 00:07:56.979 [2024-11-26T18:04:09.604Z] =================================================================================================================== 00:07:56.979 [2024-11-26T18:04:09.604Z] Total : 17939.50 70.08 0.00 0.00 0.00 0.00 0.00 00:07:56.979 00:07:57.922 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:57.922 Nvme0n1 : 7.00 17953.86 70.13 0.00 0.00 0.00 0.00 0.00 00:07:57.922 [2024-11-26T18:04:10.547Z] =================================================================================================================== 00:07:57.922 [2024-11-26T18:04:10.547Z] Total : 17953.86 70.13 0.00 0.00 0.00 0.00 0.00 00:07:57.922 00:07:58.863 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:58.863 Nvme0n1 : 8.00 17970.00 70.20 0.00 0.00 0.00 0.00 0.00 00:07:58.863 [2024-11-26T18:04:11.488Z] =================================================================================================================== 00:07:58.863 [2024-11-26T18:04:11.488Z] Total : 17970.00 70.20 0.00 0.00 0.00 0.00 0.00 00:07:58.863 00:08:00.248 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:08:00.248 Nvme0n1 : 9.00 17976.78 70.22 0.00 0.00 0.00 0.00 0.00 00:08:00.248 [2024-11-26T18:04:12.873Z] =================================================================================================================== 00:08:00.248 [2024-11-26T18:04:12.873Z] Total : 17976.78 70.22 0.00 0.00 0.00 0.00 0.00 00:08:00.248 00:08:00.824 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:08:00.824 Nvme0n1 : 10.00 17992.00 70.28 0.00 0.00 0.00 0.00 0.00 00:08:00.824 [2024-11-26T18:04:13.449Z] =================================================================================================================== 00:08:00.824 [2024-11-26T18:04:13.449Z] Total : 17992.00 70.28 0.00 0.00 0.00 0.00 0.00 00:08:00.824 00:08:01.086 00:08:01.086 Latency(us) 00:08:01.086 [2024-11-26T18:04:13.711Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:01.086 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:08:01.086 Nvme0n1 : 10.01 17993.50 70.29 0.00 0.00 7109.88 2048.00 13325.65 00:08:01.086 [2024-11-26T18:04:13.711Z] =================================================================================================================== 00:08:01.086 [2024-11-26T18:04:13.711Z] Total : 17993.50 70.29 0.00 0.00 7109.88 2048.00 13325.65 00:08:01.086 { 00:08:01.086 "results": [ 00:08:01.086 { 00:08:01.086 "job": "Nvme0n1", 00:08:01.086 "core_mask": "0x2", 00:08:01.086 "workload": "randwrite", 00:08:01.086 "status": "finished", 00:08:01.086 "queue_depth": 128, 00:08:01.086 "io_size": 4096, 00:08:01.086 "runtime": 10.006281, 00:08:01.086 "iops": 17993.49828372799, 00:08:01.086 "mibps": 70.28710267081246, 00:08:01.086 "io_failed": 0, 00:08:01.086 "io_timeout": 0, 00:08:01.086 "avg_latency_us": 7109.876572173346, 00:08:01.086 "min_latency_us": 2048.0, 00:08:01.086 "max_latency_us": 13325.653333333334 00:08:01.086 } 00:08:01.086 ], 00:08:01.086 "core_count": 1 00:08:01.086 } 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 3664115 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # '[' -z 3664115 ']' 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@958 -- # kill -0 3664115 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # uname 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3664115 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3664115' 00:08:01.086 killing process with pid 3664115 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@973 -- # kill 3664115 00:08:01.086 Received shutdown signal, test time was about 10.000000 seconds 00:08:01.086 00:08:01.086 Latency(us) 00:08:01.086 [2024-11-26T18:04:13.711Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:01.086 [2024-11-26T18:04:13.711Z] =================================================================================================================== 00:08:01.086 [2024-11-26T18:04:13.711Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@978 -- # wait 3664115 00:08:01.086 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:08:01.348 19:04:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:08:01.609 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:01.609 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:08:01.609 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:08:01.609 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:08:01.609 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 3660377 00:08:01.609 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 3660377 00:08:01.871 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 3660377 Killed "${NVMF_APP[@]}" "$@" 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@726 -- # xtrace_disable 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # nvmfpid=3666669 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # waitforlisten 3666669 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 3666669 ']' 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:01.871 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:01.871 19:04:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:08:01.871 [2024-11-26 19:04:14.303508] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:01.871 [2024-11-26 19:04:14.303561] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:01.871 [2024-11-26 19:04:14.391848] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:01.871 [2024-11-26 19:04:14.427328] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:01.871 [2024-11-26 19:04:14.427365] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:01.871 [2024-11-26 19:04:14.427373] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:01.871 [2024-11-26 19:04:14.427380] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:01.871 [2024-11-26 19:04:14.427386] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:01.871 [2024-11-26 19:04:14.427971] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@732 -- # xtrace_disable 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:08:02.813 [2024-11-26 19:04:15.303847] blobstore.c:4896:bs_recover: *NOTICE*: Performing recovery on blobstore 00:08:02.813 [2024-11-26 19:04:15.303978] blobstore.c:4843:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:08:02.813 [2024-11-26 19:04:15.304010] blobstore.c:4843:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev c4d57573-9c7f-49d0-a4e0-c1a1821acf7f 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=c4d57573-9c7f-49d0-a4e0-c1a1821acf7f 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:02.813 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:08:03.073 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b c4d57573-9c7f-49d0-a4e0-c1a1821acf7f -t 2000 00:08:03.073 [ 00:08:03.073 { 00:08:03.073 "name": "c4d57573-9c7f-49d0-a4e0-c1a1821acf7f", 00:08:03.073 "aliases": [ 00:08:03.073 "lvs/lvol" 00:08:03.073 ], 00:08:03.073 "product_name": "Logical Volume", 00:08:03.073 "block_size": 4096, 00:08:03.073 "num_blocks": 38912, 00:08:03.073 "uuid": "c4d57573-9c7f-49d0-a4e0-c1a1821acf7f", 00:08:03.073 "assigned_rate_limits": { 00:08:03.073 "rw_ios_per_sec": 0, 00:08:03.073 "rw_mbytes_per_sec": 0, 00:08:03.073 "r_mbytes_per_sec": 0, 00:08:03.073 "w_mbytes_per_sec": 0 00:08:03.073 }, 00:08:03.073 "claimed": false, 00:08:03.073 "zoned": false, 00:08:03.073 "supported_io_types": { 00:08:03.073 "read": true, 00:08:03.073 "write": true, 00:08:03.073 "unmap": true, 00:08:03.073 "flush": false, 00:08:03.073 "reset": true, 00:08:03.073 "nvme_admin": false, 00:08:03.073 "nvme_io": false, 00:08:03.073 "nvme_io_md": false, 00:08:03.073 "write_zeroes": true, 00:08:03.073 "zcopy": false, 00:08:03.073 "get_zone_info": false, 00:08:03.073 "zone_management": false, 00:08:03.073 "zone_append": false, 00:08:03.073 "compare": false, 00:08:03.073 "compare_and_write": false, 00:08:03.073 "abort": false, 00:08:03.073 "seek_hole": true, 00:08:03.073 "seek_data": true, 00:08:03.073 "copy": false, 00:08:03.073 "nvme_iov_md": false 00:08:03.073 }, 00:08:03.073 "driver_specific": { 00:08:03.073 "lvol": { 00:08:03.073 "lvol_store_uuid": "fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9", 00:08:03.073 "base_bdev": "aio_bdev", 00:08:03.073 "thin_provision": false, 00:08:03.073 "num_allocated_clusters": 38, 00:08:03.073 "snapshot": false, 00:08:03.073 "clone": false, 00:08:03.073 "esnap_clone": false 00:08:03.073 } 00:08:03.073 } 00:08:03.073 } 00:08:03.073 ] 00:08:03.073 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:08:03.073 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:03.073 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:08:03.334 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:08:03.334 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:03.334 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:08:03.594 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:08:03.594 19:04:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:08:03.594 [2024-11-26 19:04:16.119935] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # local es=0 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:08:03.594 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:03.883 request: 00:08:03.883 { 00:08:03.883 "uuid": "fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9", 00:08:03.883 "method": "bdev_lvol_get_lvstores", 00:08:03.883 "req_id": 1 00:08:03.883 } 00:08:03.883 Got JSON-RPC error response 00:08:03.883 response: 00:08:03.883 { 00:08:03.883 "code": -19, 00:08:03.883 "message": "No such device" 00:08:03.883 } 00:08:03.883 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # es=1 00:08:03.883 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:03.883 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:03.883 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:03.883 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:08:03.883 aio_bdev 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev c4d57573-9c7f-49d0-a4e0-c1a1821acf7f 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=c4d57573-9c7f-49d0-a4e0-c1a1821acf7f 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:08:04.142 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b c4d57573-9c7f-49d0-a4e0-c1a1821acf7f -t 2000 00:08:04.402 [ 00:08:04.402 { 00:08:04.402 "name": "c4d57573-9c7f-49d0-a4e0-c1a1821acf7f", 00:08:04.402 "aliases": [ 00:08:04.402 "lvs/lvol" 00:08:04.402 ], 00:08:04.402 "product_name": "Logical Volume", 00:08:04.402 "block_size": 4096, 00:08:04.402 "num_blocks": 38912, 00:08:04.402 "uuid": "c4d57573-9c7f-49d0-a4e0-c1a1821acf7f", 00:08:04.402 "assigned_rate_limits": { 00:08:04.402 "rw_ios_per_sec": 0, 00:08:04.402 "rw_mbytes_per_sec": 0, 00:08:04.402 "r_mbytes_per_sec": 0, 00:08:04.402 "w_mbytes_per_sec": 0 00:08:04.402 }, 00:08:04.402 "claimed": false, 00:08:04.402 "zoned": false, 00:08:04.402 "supported_io_types": { 00:08:04.402 "read": true, 00:08:04.402 "write": true, 00:08:04.402 "unmap": true, 00:08:04.402 "flush": false, 00:08:04.402 "reset": true, 00:08:04.402 "nvme_admin": false, 00:08:04.402 "nvme_io": false, 00:08:04.402 "nvme_io_md": false, 00:08:04.402 "write_zeroes": true, 00:08:04.402 "zcopy": false, 00:08:04.402 "get_zone_info": false, 00:08:04.402 "zone_management": false, 00:08:04.402 "zone_append": false, 00:08:04.402 "compare": false, 00:08:04.402 "compare_and_write": false, 00:08:04.402 "abort": false, 00:08:04.402 "seek_hole": true, 00:08:04.402 "seek_data": true, 00:08:04.402 "copy": false, 00:08:04.402 "nvme_iov_md": false 00:08:04.402 }, 00:08:04.402 "driver_specific": { 00:08:04.402 "lvol": { 00:08:04.402 "lvol_store_uuid": "fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9", 00:08:04.402 "base_bdev": "aio_bdev", 00:08:04.402 "thin_provision": false, 00:08:04.402 "num_allocated_clusters": 38, 00:08:04.402 "snapshot": false, 00:08:04.402 "clone": false, 00:08:04.402 "esnap_clone": false 00:08:04.402 } 00:08:04.402 } 00:08:04.402 } 00:08:04.402 ] 00:08:04.402 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:08:04.402 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:04.402 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:08:04.402 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:08:04.402 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:04.402 19:04:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:08:04.661 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:08:04.661 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete c4d57573-9c7f-49d0-a4e0-c1a1821acf7f 00:08:04.922 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u fb4e8eb9-b05e-42d8-bbd9-750d9c7d10d9 00:08:04.922 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:08:05.181 00:08:05.181 real 0m17.486s 00:08:05.181 user 0m45.662s 00:08:05.181 sys 0m2.889s 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:08:05.181 ************************************ 00:08:05.181 END TEST lvs_grow_dirty 00:08:05.181 ************************************ 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@812 -- # type=--id 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@813 -- # id=0 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@824 -- # for n in $shm_files 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:08:05.181 nvmf_trace.0 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@827 -- # return 0 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@516 -- # nvmfcleanup 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:05.181 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:05.181 rmmod nvme_tcp 00:08:05.442 rmmod nvme_fabrics 00:08:05.442 rmmod nvme_keyring 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@517 -- # '[' -n 3666669 ']' 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@518 -- # killprocess 3666669 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # '[' -z 3666669 ']' 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@958 -- # kill -0 3666669 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # uname 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3666669 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3666669' 00:08:05.442 killing process with pid 3666669 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@973 -- # kill 3666669 00:08:05.442 19:04:17 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@978 -- # wait 3666669 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-save 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-restore 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:05.442 19:04:18 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:07.989 00:08:07.989 real 0m45.288s 00:08:07.989 user 1m7.570s 00:08:07.989 sys 0m10.923s 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:08:07.989 ************************************ 00:08:07.989 END TEST nvmf_lvs_grow 00:08:07.989 ************************************ 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:08:07.989 ************************************ 00:08:07.989 START TEST nvmf_bdev_io_wait 00:08:07.989 ************************************ 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:08:07.989 * Looking for test storage... 00:08:07.989 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1693 -- # lcov --version 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:08:07.989 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:07.990 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:07.990 --rc genhtml_branch_coverage=1 00:08:07.990 --rc genhtml_function_coverage=1 00:08:07.990 --rc genhtml_legend=1 00:08:07.990 --rc geninfo_all_blocks=1 00:08:07.990 --rc geninfo_unexecuted_blocks=1 00:08:07.990 00:08:07.990 ' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:07.990 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:07.990 --rc genhtml_branch_coverage=1 00:08:07.990 --rc genhtml_function_coverage=1 00:08:07.990 --rc genhtml_legend=1 00:08:07.990 --rc geninfo_all_blocks=1 00:08:07.990 --rc geninfo_unexecuted_blocks=1 00:08:07.990 00:08:07.990 ' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:07.990 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:07.990 --rc genhtml_branch_coverage=1 00:08:07.990 --rc genhtml_function_coverage=1 00:08:07.990 --rc genhtml_legend=1 00:08:07.990 --rc geninfo_all_blocks=1 00:08:07.990 --rc geninfo_unexecuted_blocks=1 00:08:07.990 00:08:07.990 ' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:07.990 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:07.990 --rc genhtml_branch_coverage=1 00:08:07.990 --rc genhtml_function_coverage=1 00:08:07.990 --rc genhtml_legend=1 00:08:07.990 --rc geninfo_all_blocks=1 00:08:07.990 --rc geninfo_unexecuted_blocks=1 00:08:07.990 00:08:07.990 ' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:07.990 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # prepare_net_devs 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # local -g is_hw=no 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # remove_spdk_ns 00:08:07.990 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:07.991 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:07.991 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:07.991 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:08:07.991 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:08:07.991 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:08:07.991 19:04:20 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:08:16.138 Found 0000:31:00.0 (0x8086 - 0x159b) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:08:16.138 Found 0000:31:00.1 (0x8086 - 0x159b) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:08:16.138 Found net devices under 0000:31:00.0: cvl_0_0 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:08:16.138 Found net devices under 0000:31:00.1: cvl_0_1 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # is_hw=yes 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:08:16.138 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:08:16.400 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:16.400 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.690 ms 00:08:16.400 00:08:16.400 --- 10.0.0.2 ping statistics --- 00:08:16.400 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:16.400 rtt min/avg/max/mdev = 0.690/0.690/0.690/0.000 ms 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:16.400 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:16.400 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.280 ms 00:08:16.400 00:08:16.400 --- 10.0.0.1 ping statistics --- 00:08:16.400 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:16.400 rtt min/avg/max/mdev = 0.280/0.280/0.280/0.000 ms 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@450 -- # return 0 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@726 -- # xtrace_disable 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # nvmfpid=3672234 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # waitforlisten 3672234 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # '[' -z 3672234 ']' 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:16.400 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:16.400 19:04:28 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:16.661 [2024-11-26 19:04:29.049930] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:16.661 [2024-11-26 19:04:29.050004] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:16.661 [2024-11-26 19:04:29.141049] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:16.661 [2024-11-26 19:04:29.183519] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:16.661 [2024-11-26 19:04:29.183556] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:16.661 [2024-11-26 19:04:29.183565] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:16.661 [2024-11-26 19:04:29.183572] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:16.661 [2024-11-26 19:04:29.183581] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:16.661 [2024-11-26 19:04:29.185197] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:16.661 [2024-11-26 19:04:29.185325] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:16.661 [2024-11-26 19:04:29.185481] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:16.661 [2024-11-26 19:04:29.185481] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@868 -- # return 0 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@732 -- # xtrace_disable 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.606 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 [2024-11-26 19:04:29.954090] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 Malloc0 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 19:04:29 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 [2024-11-26 19:04:29.997301] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=3672532 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=3672535 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:08:17.607 { 00:08:17.607 "params": { 00:08:17.607 "name": "Nvme$subsystem", 00:08:17.607 "trtype": "$TEST_TRANSPORT", 00:08:17.607 "traddr": "$NVMF_FIRST_TARGET_IP", 00:08:17.607 "adrfam": "ipv4", 00:08:17.607 "trsvcid": "$NVMF_PORT", 00:08:17.607 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:08:17.607 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:08:17.607 "hdgst": ${hdgst:-false}, 00:08:17.607 "ddgst": ${ddgst:-false} 00:08:17.607 }, 00:08:17.607 "method": "bdev_nvme_attach_controller" 00:08:17.607 } 00:08:17.607 EOF 00:08:17.607 )") 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=3672538 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:08:17.607 { 00:08:17.607 "params": { 00:08:17.607 "name": "Nvme$subsystem", 00:08:17.607 "trtype": "$TEST_TRANSPORT", 00:08:17.607 "traddr": "$NVMF_FIRST_TARGET_IP", 00:08:17.607 "adrfam": "ipv4", 00:08:17.607 "trsvcid": "$NVMF_PORT", 00:08:17.607 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:08:17.607 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:08:17.607 "hdgst": ${hdgst:-false}, 00:08:17.607 "ddgst": ${ddgst:-false} 00:08:17.607 }, 00:08:17.607 "method": "bdev_nvme_attach_controller" 00:08:17.607 } 00:08:17.607 EOF 00:08:17.607 )") 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=3672541 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:08:17.607 { 00:08:17.607 "params": { 00:08:17.607 "name": "Nvme$subsystem", 00:08:17.607 "trtype": "$TEST_TRANSPORT", 00:08:17.607 "traddr": "$NVMF_FIRST_TARGET_IP", 00:08:17.607 "adrfam": "ipv4", 00:08:17.607 "trsvcid": "$NVMF_PORT", 00:08:17.607 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:08:17.607 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:08:17.607 "hdgst": ${hdgst:-false}, 00:08:17.607 "ddgst": ${ddgst:-false} 00:08:17.607 }, 00:08:17.607 "method": "bdev_nvme_attach_controller" 00:08:17.607 } 00:08:17.607 EOF 00:08:17.607 )") 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:08:17.607 { 00:08:17.607 "params": { 00:08:17.607 "name": "Nvme$subsystem", 00:08:17.607 "trtype": "$TEST_TRANSPORT", 00:08:17.607 "traddr": "$NVMF_FIRST_TARGET_IP", 00:08:17.607 "adrfam": "ipv4", 00:08:17.607 "trsvcid": "$NVMF_PORT", 00:08:17.607 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:08:17.607 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:08:17.607 "hdgst": ${hdgst:-false}, 00:08:17.607 "ddgst": ${ddgst:-false} 00:08:17.607 }, 00:08:17.607 "method": "bdev_nvme_attach_controller" 00:08:17.607 } 00:08:17.607 EOF 00:08:17.607 )") 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 3672532 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:08:17.607 "params": { 00:08:17.607 "name": "Nvme1", 00:08:17.607 "trtype": "tcp", 00:08:17.607 "traddr": "10.0.0.2", 00:08:17.607 "adrfam": "ipv4", 00:08:17.607 "trsvcid": "4420", 00:08:17.607 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:08:17.607 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:08:17.607 "hdgst": false, 00:08:17.607 "ddgst": false 00:08:17.607 }, 00:08:17.607 "method": "bdev_nvme_attach_controller" 00:08:17.607 }' 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:08:17.607 "params": { 00:08:17.607 "name": "Nvme1", 00:08:17.607 "trtype": "tcp", 00:08:17.607 "traddr": "10.0.0.2", 00:08:17.607 "adrfam": "ipv4", 00:08:17.607 "trsvcid": "4420", 00:08:17.607 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:08:17.607 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:08:17.607 "hdgst": false, 00:08:17.607 "ddgst": false 00:08:17.607 }, 00:08:17.607 "method": "bdev_nvme_attach_controller" 00:08:17.607 }' 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:08:17.607 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:08:17.607 "params": { 00:08:17.607 "name": "Nvme1", 00:08:17.607 "trtype": "tcp", 00:08:17.607 "traddr": "10.0.0.2", 00:08:17.607 "adrfam": "ipv4", 00:08:17.608 "trsvcid": "4420", 00:08:17.608 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:08:17.608 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:08:17.608 "hdgst": false, 00:08:17.608 "ddgst": false 00:08:17.608 }, 00:08:17.608 "method": "bdev_nvme_attach_controller" 00:08:17.608 }' 00:08:17.608 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:08:17.608 19:04:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:08:17.608 "params": { 00:08:17.608 "name": "Nvme1", 00:08:17.608 "trtype": "tcp", 00:08:17.608 "traddr": "10.0.0.2", 00:08:17.608 "adrfam": "ipv4", 00:08:17.608 "trsvcid": "4420", 00:08:17.608 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:08:17.608 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:08:17.608 "hdgst": false, 00:08:17.608 "ddgst": false 00:08:17.608 }, 00:08:17.608 "method": "bdev_nvme_attach_controller" 00:08:17.608 }' 00:08:17.608 [2024-11-26 19:04:30.054444] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:17.608 [2024-11-26 19:04:30.054499] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:08:17.608 [2024-11-26 19:04:30.055013] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:17.608 [2024-11-26 19:04:30.055058] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:08:17.608 [2024-11-26 19:04:30.055695] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:17.608 [2024-11-26 19:04:30.055745] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:08:17.608 [2024-11-26 19:04:30.056012] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:17.608 [2024-11-26 19:04:30.056058] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:08:17.869 [2024-11-26 19:04:30.228723] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.869 [2024-11-26 19:04:30.258591] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:08:17.869 [2024-11-26 19:04:30.273277] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.869 [2024-11-26 19:04:30.301602] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:08:17.869 [2024-11-26 19:04:30.320442] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.869 [2024-11-26 19:04:30.348824] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:08:17.869 [2024-11-26 19:04:30.382416] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.869 [2024-11-26 19:04:30.411471] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:08:17.869 Running I/O for 1 seconds... 00:08:17.869 Running I/O for 1 seconds... 00:08:18.131 Running I/O for 1 seconds... 00:08:18.131 Running I/O for 1 seconds... 00:08:19.076 18828.00 IOPS, 73.55 MiB/s 00:08:19.076 Latency(us) 00:08:19.076 [2024-11-26T18:04:31.701Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:19.076 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:08:19.076 Nvme1n1 : 1.01 18882.50 73.76 0.00 0.00 6759.95 3795.63 15619.41 00:08:19.076 [2024-11-26T18:04:31.701Z] =================================================================================================================== 00:08:19.076 [2024-11-26T18:04:31.701Z] Total : 18882.50 73.76 0.00 0.00 6759.95 3795.63 15619.41 00:08:19.076 183320.00 IOPS, 716.09 MiB/s 00:08:19.076 Latency(us) 00:08:19.076 [2024-11-26T18:04:31.701Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:19.076 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:08:19.076 Nvme1n1 : 1.00 182957.69 714.68 0.00 0.00 695.19 293.55 1952.43 00:08:19.076 [2024-11-26T18:04:31.701Z] =================================================================================================================== 00:08:19.076 [2024-11-26T18:04:31.701Z] Total : 182957.69 714.68 0.00 0.00 695.19 293.55 1952.43 00:08:19.076 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 3672535 00:08:19.076 12709.00 IOPS, 49.64 MiB/s 00:08:19.076 Latency(us) 00:08:19.076 [2024-11-26T18:04:31.701Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:19.076 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:08:19.076 Nvme1n1 : 1.01 12777.59 49.91 0.00 0.00 9988.86 3713.71 22719.15 00:08:19.076 [2024-11-26T18:04:31.701Z] =================================================================================================================== 00:08:19.076 [2024-11-26T18:04:31.701Z] Total : 12777.59 49.91 0.00 0.00 9988.86 3713.71 22719.15 00:08:19.076 10826.00 IOPS, 42.29 MiB/s 00:08:19.076 Latency(us) 00:08:19.076 [2024-11-26T18:04:31.701Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:19.076 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:08:19.076 Nvme1n1 : 1.01 10879.68 42.50 0.00 0.00 11720.94 5297.49 23156.05 00:08:19.076 [2024-11-26T18:04:31.701Z] =================================================================================================================== 00:08:19.076 [2024-11-26T18:04:31.701Z] Total : 10879.68 42.50 0.00 0.00 11720.94 5297.49 23156.05 00:08:19.076 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 3672538 00:08:19.076 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 3672541 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # nvmfcleanup 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:19.338 rmmod nvme_tcp 00:08:19.338 rmmod nvme_fabrics 00:08:19.338 rmmod nvme_keyring 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@517 -- # '[' -n 3672234 ']' 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # killprocess 3672234 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # '[' -z 3672234 ']' 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@958 -- # kill -0 3672234 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # uname 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3672234 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3672234' 00:08:19.338 killing process with pid 3672234 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@973 -- # kill 3672234 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@978 -- # wait 3672234 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-save 00:08:19.338 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-restore 00:08:19.600 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:19.600 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:19.600 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:19.600 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:19.600 19:04:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:21.515 00:08:21.515 real 0m13.829s 00:08:21.515 user 0m18.623s 00:08:21.515 sys 0m7.844s 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:08:21.515 ************************************ 00:08:21.515 END TEST nvmf_bdev_io_wait 00:08:21.515 ************************************ 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:08:21.515 ************************************ 00:08:21.515 START TEST nvmf_queue_depth 00:08:21.515 ************************************ 00:08:21.515 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:08:21.777 * Looking for test storage... 00:08:21.777 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1693 -- # lcov --version 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:08:21.777 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:21.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:21.778 --rc genhtml_branch_coverage=1 00:08:21.778 --rc genhtml_function_coverage=1 00:08:21.778 --rc genhtml_legend=1 00:08:21.778 --rc geninfo_all_blocks=1 00:08:21.778 --rc geninfo_unexecuted_blocks=1 00:08:21.778 00:08:21.778 ' 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:21.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:21.778 --rc genhtml_branch_coverage=1 00:08:21.778 --rc genhtml_function_coverage=1 00:08:21.778 --rc genhtml_legend=1 00:08:21.778 --rc geninfo_all_blocks=1 00:08:21.778 --rc geninfo_unexecuted_blocks=1 00:08:21.778 00:08:21.778 ' 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:21.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:21.778 --rc genhtml_branch_coverage=1 00:08:21.778 --rc genhtml_function_coverage=1 00:08:21.778 --rc genhtml_legend=1 00:08:21.778 --rc geninfo_all_blocks=1 00:08:21.778 --rc geninfo_unexecuted_blocks=1 00:08:21.778 00:08:21.778 ' 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:21.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:21.778 --rc genhtml_branch_coverage=1 00:08:21.778 --rc genhtml_function_coverage=1 00:08:21.778 --rc genhtml_legend=1 00:08:21.778 --rc geninfo_all_blocks=1 00:08:21.778 --rc geninfo_unexecuted_blocks=1 00:08:21.778 00:08:21.778 ' 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:21.778 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:21.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@476 -- # prepare_net_devs 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@438 -- # local -g is_hw=no 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # remove_spdk_ns 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:08:21.779 19:04:34 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:08:29.925 Found 0000:31:00.0 (0x8086 - 0x159b) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:08:29.925 Found 0000:31:00.1 (0x8086 - 0x159b) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:29.925 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:08:29.926 Found net devices under 0000:31:00.0: cvl_0_0 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:08:29.926 Found net devices under 0000:31:00.1: cvl_0_1 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # is_hw=yes 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:08:29.926 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:08:30.186 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:30.186 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.496 ms 00:08:30.186 00:08:30.186 --- 10.0.0.2 ping statistics --- 00:08:30.186 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:30.186 rtt min/avg/max/mdev = 0.496/0.496/0.496/0.000 ms 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:30.186 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:30.186 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.306 ms 00:08:30.186 00:08:30.186 --- 10.0.0.1 ping statistics --- 00:08:30.186 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:30.186 rtt min/avg/max/mdev = 0.306/0.306/0.306/0.000 ms 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@450 -- # return 0 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@726 -- # xtrace_disable 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:30.186 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@509 -- # nvmfpid=3677644 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@510 -- # waitforlisten 3677644 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 3677644 ']' 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:30.448 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:30.448 19:04:42 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:30.448 [2024-11-26 19:04:42.869927] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:30.448 [2024-11-26 19:04:42.870016] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:30.448 [2024-11-26 19:04:42.984765] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:30.448 [2024-11-26 19:04:43.035053] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:30.448 [2024-11-26 19:04:43.035112] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:30.448 [2024-11-26 19:04:43.035121] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:30.448 [2024-11-26 19:04:43.035128] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:30.448 [2024-11-26 19:04:43.035135] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:30.448 [2024-11-26 19:04:43.036030] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@732 -- # xtrace_disable 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:31.391 [2024-11-26 19:04:43.734994] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:31.391 Malloc0 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:31.391 [2024-11-26 19:04:43.796395] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=3677895 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 3677895 /var/tmp/bdevperf.sock 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 3677895 ']' 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:08:31.391 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:31.391 19:04:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:31.391 [2024-11-26 19:04:43.854563] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:08:31.391 [2024-11-26 19:04:43.854631] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3677895 ] 00:08:31.391 [2024-11-26 19:04:43.938524] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:31.391 [2024-11-26 19:04:43.980700] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:32.333 19:04:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:32.333 19:04:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:08:32.333 19:04:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:08:32.333 19:04:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.333 19:04:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:32.333 NVMe0n1 00:08:32.333 19:04:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.333 19:04:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:08:32.333 Running I/O for 10 seconds... 00:08:34.221 8888.00 IOPS, 34.72 MiB/s [2024-11-26T18:04:48.234Z] 10240.00 IOPS, 40.00 MiB/s [2024-11-26T18:04:49.176Z] 10691.00 IOPS, 41.76 MiB/s [2024-11-26T18:04:50.131Z] 10944.00 IOPS, 42.75 MiB/s [2024-11-26T18:04:51.200Z] 11063.20 IOPS, 43.22 MiB/s [2024-11-26T18:04:52.144Z] 11215.00 IOPS, 43.81 MiB/s [2024-11-26T18:04:53.088Z] 11262.57 IOPS, 43.99 MiB/s [2024-11-26T18:04:54.028Z] 11310.88 IOPS, 44.18 MiB/s [2024-11-26T18:04:54.968Z] 11349.00 IOPS, 44.33 MiB/s [2024-11-26T18:04:54.968Z] 11366.70 IOPS, 44.40 MiB/s 00:08:42.343 Latency(us) 00:08:42.343 [2024-11-26T18:04:54.968Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:42.343 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:08:42.343 Verification LBA range: start 0x0 length 0x4000 00:08:42.343 NVMe0n1 : 10.06 11397.59 44.52 0.00 0.00 89555.33 24466.77 77769.39 00:08:42.343 [2024-11-26T18:04:54.968Z] =================================================================================================================== 00:08:42.343 [2024-11-26T18:04:54.968Z] Total : 11397.59 44.52 0.00 0.00 89555.33 24466.77 77769.39 00:08:42.343 { 00:08:42.343 "results": [ 00:08:42.343 { 00:08:42.343 "job": "NVMe0n1", 00:08:42.343 "core_mask": "0x1", 00:08:42.343 "workload": "verify", 00:08:42.343 "status": "finished", 00:08:42.343 "verify_range": { 00:08:42.343 "start": 0, 00:08:42.343 "length": 16384 00:08:42.343 }, 00:08:42.343 "queue_depth": 1024, 00:08:42.343 "io_size": 4096, 00:08:42.343 "runtime": 10.061605, 00:08:42.343 "iops": 11397.585176520048, 00:08:42.343 "mibps": 44.52181709578144, 00:08:42.343 "io_failed": 0, 00:08:42.343 "io_timeout": 0, 00:08:42.343 "avg_latency_us": 89555.33050989146, 00:08:42.343 "min_latency_us": 24466.773333333334, 00:08:42.343 "max_latency_us": 77769.38666666667 00:08:42.343 } 00:08:42.343 ], 00:08:42.343 "core_count": 1 00:08:42.343 } 00:08:42.343 19:04:54 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 3677895 00:08:42.343 19:04:54 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 3677895 ']' 00:08:42.343 19:04:54 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 3677895 00:08:42.343 19:04:54 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:08:42.343 19:04:54 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:42.343 19:04:54 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3677895 00:08:42.602 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:42.602 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:42.602 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3677895' 00:08:42.602 killing process with pid 3677895 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 3677895 00:08:42.603 Received shutdown signal, test time was about 10.000000 seconds 00:08:42.603 00:08:42.603 Latency(us) 00:08:42.603 [2024-11-26T18:04:55.228Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:42.603 [2024-11-26T18:04:55.228Z] =================================================================================================================== 00:08:42.603 [2024-11-26T18:04:55.228Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 3677895 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@516 -- # nvmfcleanup 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:42.603 rmmod nvme_tcp 00:08:42.603 rmmod nvme_fabrics 00:08:42.603 rmmod nvme_keyring 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@517 -- # '[' -n 3677644 ']' 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@518 -- # killprocess 3677644 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 3677644 ']' 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 3677644 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:42.603 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3677644 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3677644' 00:08:42.862 killing process with pid 3677644 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 3677644 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 3677644 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:08:42.862 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-restore 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-save 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:42.863 19:04:55 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:45.405 00:08:45.405 real 0m23.333s 00:08:45.405 user 0m25.949s 00:08:45.405 sys 0m7.628s 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:08:45.405 ************************************ 00:08:45.405 END TEST nvmf_queue_depth 00:08:45.405 ************************************ 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:08:45.405 ************************************ 00:08:45.405 START TEST nvmf_target_multipath 00:08:45.405 ************************************ 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:08:45.405 * Looking for test storage... 00:08:45.405 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1693 -- # lcov --version 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:45.405 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:45.405 --rc genhtml_branch_coverage=1 00:08:45.405 --rc genhtml_function_coverage=1 00:08:45.405 --rc genhtml_legend=1 00:08:45.405 --rc geninfo_all_blocks=1 00:08:45.405 --rc geninfo_unexecuted_blocks=1 00:08:45.405 00:08:45.405 ' 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:45.405 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:45.405 --rc genhtml_branch_coverage=1 00:08:45.405 --rc genhtml_function_coverage=1 00:08:45.405 --rc genhtml_legend=1 00:08:45.405 --rc geninfo_all_blocks=1 00:08:45.405 --rc geninfo_unexecuted_blocks=1 00:08:45.405 00:08:45.405 ' 00:08:45.405 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:45.406 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:45.406 --rc genhtml_branch_coverage=1 00:08:45.406 --rc genhtml_function_coverage=1 00:08:45.406 --rc genhtml_legend=1 00:08:45.406 --rc geninfo_all_blocks=1 00:08:45.406 --rc geninfo_unexecuted_blocks=1 00:08:45.406 00:08:45.406 ' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:45.406 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:45.406 --rc genhtml_branch_coverage=1 00:08:45.406 --rc genhtml_function_coverage=1 00:08:45.406 --rc genhtml_legend=1 00:08:45.406 --rc geninfo_all_blocks=1 00:08:45.406 --rc geninfo_unexecuted_blocks=1 00:08:45.406 00:08:45.406 ' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:45.406 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@476 -- # prepare_net_devs 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@438 -- # local -g is_hw=no 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # remove_spdk_ns 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:08:45.406 19:04:57 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:08:53.553 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:53.553 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:08:53.553 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:08:53.553 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:08:53.553 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:08:53.554 Found 0000:31:00.0 (0x8086 - 0x159b) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:08:53.554 Found 0000:31:00.1 (0x8086 - 0x159b) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:08:53.554 Found net devices under 0000:31:00.0: cvl_0_0 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:08:53.554 Found net devices under 0000:31:00.1: cvl_0_1 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # is_hw=yes 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:53.554 19:05:05 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:08:53.554 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:53.554 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.701 ms 00:08:53.554 00:08:53.554 --- 10.0.0.2 ping statistics --- 00:08:53.554 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:53.554 rtt min/avg/max/mdev = 0.701/0.701/0.701/0.000 ms 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:53.554 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:53.554 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.289 ms 00:08:53.554 00:08:53.554 --- 10.0.0.1 ping statistics --- 00:08:53.554 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:53.554 rtt min/avg/max/mdev = 0.289/0.289/0.289/0.000 ms 00:08:53.554 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@450 -- # return 0 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:08:53.555 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:08:53.815 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:08:53.815 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:08:53.816 only one NIC for nvmf test 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:53.816 rmmod nvme_tcp 00:08:53.816 rmmod nvme_fabrics 00:08:53.816 rmmod nvme_keyring 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:53.816 19:05:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:55.728 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:55.728 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:08:55.728 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:08:55.728 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:08:55.728 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:55.989 00:08:55.989 real 0m10.857s 00:08:55.989 user 0m2.323s 00:08:55.989 sys 0m6.455s 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:08:55.989 ************************************ 00:08:55.989 END TEST nvmf_target_multipath 00:08:55.989 ************************************ 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:08:55.989 ************************************ 00:08:55.989 START TEST nvmf_zcopy 00:08:55.989 ************************************ 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:08:55.989 * Looking for test storage... 00:08:55.989 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1693 -- # lcov --version 00:08:55.989 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:56.250 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.250 --rc genhtml_branch_coverage=1 00:08:56.250 --rc genhtml_function_coverage=1 00:08:56.250 --rc genhtml_legend=1 00:08:56.250 --rc geninfo_all_blocks=1 00:08:56.250 --rc geninfo_unexecuted_blocks=1 00:08:56.250 00:08:56.250 ' 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:56.250 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.250 --rc genhtml_branch_coverage=1 00:08:56.250 --rc genhtml_function_coverage=1 00:08:56.250 --rc genhtml_legend=1 00:08:56.250 --rc geninfo_all_blocks=1 00:08:56.250 --rc geninfo_unexecuted_blocks=1 00:08:56.250 00:08:56.250 ' 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:56.250 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.250 --rc genhtml_branch_coverage=1 00:08:56.250 --rc genhtml_function_coverage=1 00:08:56.250 --rc genhtml_legend=1 00:08:56.250 --rc geninfo_all_blocks=1 00:08:56.250 --rc geninfo_unexecuted_blocks=1 00:08:56.250 00:08:56.250 ' 00:08:56.250 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:56.250 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.250 --rc genhtml_branch_coverage=1 00:08:56.250 --rc genhtml_function_coverage=1 00:08:56.250 --rc genhtml_legend=1 00:08:56.250 --rc geninfo_all_blocks=1 00:08:56.250 --rc geninfo_unexecuted_blocks=1 00:08:56.250 00:08:56.250 ' 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:56.251 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@476 -- # prepare_net_devs 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@438 -- # local -g is_hw=no 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # remove_spdk_ns 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:08:56.251 19:05:08 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:04.395 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:09:04.396 Found 0000:31:00.0 (0x8086 - 0x159b) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:09:04.396 Found 0000:31:00.1 (0x8086 - 0x159b) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:09:04.396 Found net devices under 0000:31:00.0: cvl_0_0 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:09:04.396 Found net devices under 0000:31:00.1: cvl_0_1 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # is_hw=yes 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:04.396 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:04.396 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.678 ms 00:09:04.396 00:09:04.396 --- 10.0.0.2 ping statistics --- 00:09:04.396 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:04.396 rtt min/avg/max/mdev = 0.678/0.678/0.678/0.000 ms 00:09:04.396 19:05:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:04.396 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:04.396 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.337 ms 00:09:04.396 00:09:04.396 --- 10.0.0.1 ping statistics --- 00:09:04.396 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:04.396 rtt min/avg/max/mdev = 0.337/0.337/0.337/0.000 ms 00:09:04.396 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:04.396 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@450 -- # return 0 00:09:04.396 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:09:04.396 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:04.397 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:09:04.397 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:09:04.397 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:04.397 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:09:04.397 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@509 -- # nvmfpid=3689739 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@510 -- # waitforlisten 3689739 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@835 -- # '[' -z 3689739 ']' 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:04.658 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:04.658 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:04.658 [2024-11-26 19:05:17.115677] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:09:04.658 [2024-11-26 19:05:17.115745] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:04.658 [2024-11-26 19:05:17.224022] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:04.658 [2024-11-26 19:05:17.273290] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:04.658 [2024-11-26 19:05:17.273344] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:04.658 [2024-11-26 19:05:17.273353] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:04.658 [2024-11-26 19:05:17.273360] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:04.658 [2024-11-26 19:05:17.273366] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:04.658 [2024-11-26 19:05:17.274181] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@868 -- # return 0 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:05.600 [2024-11-26 19:05:17.984667] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.600 19:05:17 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:05.600 [2024-11-26 19:05:18.008947] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:05.600 malloc0 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:09:05.600 { 00:09:05.600 "params": { 00:09:05.600 "name": "Nvme$subsystem", 00:09:05.600 "trtype": "$TEST_TRANSPORT", 00:09:05.600 "traddr": "$NVMF_FIRST_TARGET_IP", 00:09:05.600 "adrfam": "ipv4", 00:09:05.600 "trsvcid": "$NVMF_PORT", 00:09:05.600 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:09:05.600 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:09:05.600 "hdgst": ${hdgst:-false}, 00:09:05.600 "ddgst": ${ddgst:-false} 00:09:05.600 }, 00:09:05.600 "method": "bdev_nvme_attach_controller" 00:09:05.600 } 00:09:05.600 EOF 00:09:05.600 )") 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:09:05.600 19:05:18 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:09:05.600 "params": { 00:09:05.600 "name": "Nvme1", 00:09:05.600 "trtype": "tcp", 00:09:05.600 "traddr": "10.0.0.2", 00:09:05.600 "adrfam": "ipv4", 00:09:05.600 "trsvcid": "4420", 00:09:05.600 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:09:05.600 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:09:05.600 "hdgst": false, 00:09:05.600 "ddgst": false 00:09:05.600 }, 00:09:05.600 "method": "bdev_nvme_attach_controller" 00:09:05.600 }' 00:09:05.600 [2024-11-26 19:05:18.109685] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:09:05.600 [2024-11-26 19:05:18.109750] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3690036 ] 00:09:05.600 [2024-11-26 19:05:18.192209] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:05.860 [2024-11-26 19:05:18.233597] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:05.860 Running I/O for 10 seconds... 00:09:07.817 6647.00 IOPS, 51.93 MiB/s [2024-11-26T18:05:21.828Z] 7416.50 IOPS, 57.94 MiB/s [2024-11-26T18:05:22.772Z] 8185.67 IOPS, 63.95 MiB/s [2024-11-26T18:05:23.717Z] 8576.00 IOPS, 67.00 MiB/s [2024-11-26T18:05:24.661Z] 8806.60 IOPS, 68.80 MiB/s [2024-11-26T18:05:25.605Z] 8959.00 IOPS, 69.99 MiB/s [2024-11-26T18:05:26.548Z] 9072.14 IOPS, 70.88 MiB/s [2024-11-26T18:05:27.491Z] 9157.88 IOPS, 71.55 MiB/s [2024-11-26T18:05:28.878Z] 9221.89 IOPS, 72.05 MiB/s [2024-11-26T18:05:28.878Z] 9272.30 IOPS, 72.44 MiB/s 00:09:16.253 Latency(us) 00:09:16.253 [2024-11-26T18:05:28.878Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:09:16.253 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:09:16.253 Verification LBA range: start 0x0 length 0x1000 00:09:16.253 Nvme1n1 : 10.01 9274.71 72.46 0.00 0.00 13748.70 1993.39 27743.57 00:09:16.253 [2024-11-26T18:05:28.878Z] =================================================================================================================== 00:09:16.253 [2024-11-26T18:05:28.878Z] Total : 9274.71 72.46 0.00 0.00 13748.70 1993.39 27743.57 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=3692107 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:09:16.253 { 00:09:16.253 "params": { 00:09:16.253 "name": "Nvme$subsystem", 00:09:16.253 "trtype": "$TEST_TRANSPORT", 00:09:16.253 "traddr": "$NVMF_FIRST_TARGET_IP", 00:09:16.253 "adrfam": "ipv4", 00:09:16.253 "trsvcid": "$NVMF_PORT", 00:09:16.253 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:09:16.253 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:09:16.253 "hdgst": ${hdgst:-false}, 00:09:16.253 "ddgst": ${ddgst:-false} 00:09:16.253 }, 00:09:16.253 "method": "bdev_nvme_attach_controller" 00:09:16.253 } 00:09:16.253 EOF 00:09:16.253 )") 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:09:16.253 [2024-11-26 19:05:28.584208] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.253 [2024-11-26 19:05:28.584237] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:09:16.253 19:05:28 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:09:16.253 "params": { 00:09:16.253 "name": "Nvme1", 00:09:16.253 "trtype": "tcp", 00:09:16.253 "traddr": "10.0.0.2", 00:09:16.253 "adrfam": "ipv4", 00:09:16.254 "trsvcid": "4420", 00:09:16.254 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:09:16.254 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:09:16.254 "hdgst": false, 00:09:16.254 "ddgst": false 00:09:16.254 }, 00:09:16.254 "method": "bdev_nvme_attach_controller" 00:09:16.254 }' 00:09:16.254 [2024-11-26 19:05:28.596206] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.596215] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.608235] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.608243] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.620265] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.620277] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.629146] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:09:16.254 [2024-11-26 19:05:28.629194] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3692107 ] 00:09:16.254 [2024-11-26 19:05:28.632295] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.632303] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.644326] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.644334] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.656356] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.656364] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.668388] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.668396] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.680419] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.680426] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.692450] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.692458] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.704481] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.704489] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.705705] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:16.254 [2024-11-26 19:05:28.716512] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.716521] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.728543] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.728552] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.740572] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.740583] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.740746] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:16.254 [2024-11-26 19:05:28.752607] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.752616] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.764640] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.764653] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.776661] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.776672] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.788692] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.788700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.800724] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.800731] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.812751] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.812759] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.824796] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.824813] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.836815] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.836825] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.848844] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.848853] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.860877] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.860885] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.254 [2024-11-26 19:05:28.872912] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.254 [2024-11-26 19:05:28.872920] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:28.884940] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:28.884950] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:28.896974] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:28.896986] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:28.909005] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:28.909015] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:28.958384] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:28.958399] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:28.969163] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:28.969173] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 Running I/O for 5 seconds... 00:09:16.515 [2024-11-26 19:05:28.984040] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:28.984056] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:28.997266] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:28.997283] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.010829] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.010845] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.024180] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.024196] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.037546] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.037562] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.050005] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.050021] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.062508] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.062523] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.076085] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.076100] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.088908] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.088928] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.102174] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.102190] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.115061] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.115077] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.515 [2024-11-26 19:05:29.127837] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.515 [2024-11-26 19:05:29.127853] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.774 [2024-11-26 19:05:29.140721] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.774 [2024-11-26 19:05:29.140737] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.774 [2024-11-26 19:05:29.154319] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.774 [2024-11-26 19:05:29.154335] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.774 [2024-11-26 19:05:29.167869] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.774 [2024-11-26 19:05:29.167884] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.774 [2024-11-26 19:05:29.181301] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.774 [2024-11-26 19:05:29.181317] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.774 [2024-11-26 19:05:29.194793] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.774 [2024-11-26 19:05:29.194808] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.774 [2024-11-26 19:05:29.208050] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.208065] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.221239] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.221254] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.233990] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.234006] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.247574] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.247589] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.261002] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.261017] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.273426] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.273441] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.286434] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.286450] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.299452] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.299467] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.312246] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.312261] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.324998] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.325013] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.337657] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.337677] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.350028] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.350044] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.363660] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.363675] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.376442] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.376460] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:16.775 [2024-11-26 19:05:29.389495] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:16.775 [2024-11-26 19:05:29.389511] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.402806] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.402823] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.415490] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.415506] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.428446] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.428462] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.441372] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.441388] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.453922] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.453938] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.467130] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.467146] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.480521] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.480537] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.494048] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.494064] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.506281] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.506297] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.519111] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.519127] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.532800] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.532816] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.545945] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.545962] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.559370] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.559386] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.572731] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.572747] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.586136] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.586155] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.599434] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.599450] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.613038] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.613053] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.625688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.625703] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.638348] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.638363] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.035 [2024-11-26 19:05:29.651642] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.035 [2024-11-26 19:05:29.651658] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.664955] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.664971] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.677640] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.677656] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.689908] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.689923] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.703184] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.703201] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.716623] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.716638] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.730086] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.730101] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.742840] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.742855] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.756344] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.756359] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.769929] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.769945] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.782973] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.782989] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.796376] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.796392] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.809312] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.809327] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.822769] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.822785] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.835972] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.835995] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.849568] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.849584] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.862331] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.862347] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.875750] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.875764] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.889103] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.889119] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.902009] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.902024] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.296 [2024-11-26 19:05:29.914625] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.296 [2024-11-26 19:05:29.914640] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:29.927266] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:29.927282] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:29.941212] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:29.941228] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:29.954526] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:29.954542] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:29.967703] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:29.967719] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 19058.00 IOPS, 148.89 MiB/s [2024-11-26T18:05:30.182Z] [2024-11-26 19:05:29.981266] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:29.981281] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:29.994975] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:29.994991] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.007972] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.007990] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.020987] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.021005] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.033701] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.033717] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.046972] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.046990] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.059470] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.059486] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.071742] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.071758] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.085711] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.085733] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.097174] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.097194] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.113453] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.113474] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.129340] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.129362] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.143949] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.143967] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.157310] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.157327] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.557 [2024-11-26 19:05:30.170853] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.557 [2024-11-26 19:05:30.170874] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.184306] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.184323] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.196967] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.196983] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.210495] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.210510] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.223841] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.223856] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.236789] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.236804] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.249680] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.249696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.262520] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.262535] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.275619] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.275634] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.288913] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.288929] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.301472] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.301487] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.314273] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.314288] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.327345] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.327360] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.340910] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.340926] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.353606] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.353621] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.366243] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.366258] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.378984] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.379000] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.391253] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.391268] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.404841] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.404858] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.417758] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.417774] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:17.818 [2024-11-26 19:05:30.431530] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:17.818 [2024-11-26 19:05:30.431545] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.444018] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.444033] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.457459] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.457474] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.470897] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.470913] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.484228] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.484243] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.497917] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.497932] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.510717] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.510733] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.524069] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.524084] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.537728] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.537743] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.551050] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.551065] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.564697] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.564712] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.577664] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.577683] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.591213] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.591228] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.603778] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.603793] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.616273] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.616289] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.629860] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.629883] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.078 [2024-11-26 19:05:30.643275] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.078 [2024-11-26 19:05:30.643291] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.079 [2024-11-26 19:05:30.656109] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.079 [2024-11-26 19:05:30.656125] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.079 [2024-11-26 19:05:30.669561] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.079 [2024-11-26 19:05:30.669576] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.079 [2024-11-26 19:05:30.682785] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.079 [2024-11-26 19:05:30.682799] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.079 [2024-11-26 19:05:30.695532] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.079 [2024-11-26 19:05:30.695547] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.338 [2024-11-26 19:05:30.708742] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.338 [2024-11-26 19:05:30.708758] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.338 [2024-11-26 19:05:30.722263] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.338 [2024-11-26 19:05:30.722280] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.338 [2024-11-26 19:05:30.735007] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.338 [2024-11-26 19:05:30.735023] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.338 [2024-11-26 19:05:30.747732] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.338 [2024-11-26 19:05:30.747748] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.760849] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.760869] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.774302] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.774316] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.787268] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.787283] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.800203] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.800218] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.813082] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.813097] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.826569] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.826589] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.839877] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.839892] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.853046] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.853061] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.865791] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.865805] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.879218] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.879233] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.892541] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.892557] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.906027] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.906042] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.919702] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.919717] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.932345] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.932360] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.945730] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.945746] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.339 [2024-11-26 19:05:30.959001] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.339 [2024-11-26 19:05:30.959016] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:30.972694] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:30.972710] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 19008.00 IOPS, 148.50 MiB/s [2024-11-26T18:05:31.224Z] [2024-11-26 19:05:30.985548] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:30.985563] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:30.998184] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:30.998200] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.010892] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.010908] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.024554] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.024569] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.037399] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.037414] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.050255] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.050271] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.063503] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.063517] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.076783] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.076803] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.090239] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.090254] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.102928] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.102944] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.115695] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.115711] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.128626] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.128641] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.141655] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.141671] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.154564] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.154580] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.168138] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.168156] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.180449] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.180465] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.193751] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.193766] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.206516] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.206531] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.599 [2024-11-26 19:05:31.219508] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.599 [2024-11-26 19:05:31.219523] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.232756] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.232772] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.245483] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.245499] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.258135] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.258151] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.271598] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.271614] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.284902] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.284917] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.298279] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.298294] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.311308] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.311324] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.324405] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.324421] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.337743] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.337759] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.351159] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.351175] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.364383] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.364399] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.377133] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.377148] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.389645] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.389661] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.403310] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.403325] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.416450] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.416465] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.429404] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.429419] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.442203] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.442219] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.455504] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.455520] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.468638] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.468653] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:18.861 [2024-11-26 19:05:31.481371] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:18.861 [2024-11-26 19:05:31.481387] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.494327] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.494343] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.506881] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.506897] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.519288] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.519305] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.532719] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.532735] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.545913] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.545929] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.558522] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.558537] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.571624] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.571639] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.585189] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.585205] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.598195] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.598210] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.611453] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.611468] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.624959] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.624976] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.637783] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.637799] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.650546] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.650562] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.663104] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.663120] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.675473] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.675488] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.688918] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.688934] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.701813] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.701829] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.714576] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.714592] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.728112] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.728128] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.124 [2024-11-26 19:05:31.740455] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.124 [2024-11-26 19:05:31.740471] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.753913] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.753929] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.767318] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.767333] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.779894] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.779910] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.792488] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.792503] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.806021] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.806037] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.818399] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.818414] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.831112] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.831127] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.844292] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.844308] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.857237] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.857252] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.870661] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.870676] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.884248] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.884264] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.896996] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.897012] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.909915] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.909931] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.922719] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.922735] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.936153] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.936169] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.949783] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.949799] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.962367] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.962382] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.385 [2024-11-26 19:05:31.975210] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.385 [2024-11-26 19:05:31.975225] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.386 19098.00 IOPS, 149.20 MiB/s [2024-11-26T18:05:32.011Z] [2024-11-26 19:05:31.988131] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.386 [2024-11-26 19:05:31.988147] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.386 [2024-11-26 19:05:32.000759] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.386 [2024-11-26 19:05:32.000774] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.014236] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.014251] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.027671] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.027687] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.040728] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.040744] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.054012] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.054031] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.067471] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.067486] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.081185] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.081201] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.094899] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.094914] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.107322] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.107338] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.120845] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.120860] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.133670] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.133685] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.147033] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.147048] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.159875] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.159891] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.172995] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.173010] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.186535] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.186550] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.199634] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.199649] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.212643] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.212658] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.226153] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.226168] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.238513] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.238528] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.250909] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.250925] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.647 [2024-11-26 19:05:32.263844] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.647 [2024-11-26 19:05:32.263859] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.276395] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.276410] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.289727] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.289742] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.302612] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.302631] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.316119] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.316135] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.328936] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.328951] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.342315] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.342330] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.355680] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.355695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.369441] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.369456] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.381962] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.381976] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.395071] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.395085] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.407860] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.407879] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.420722] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.420738] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.908 [2024-11-26 19:05:32.434159] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.908 [2024-11-26 19:05:32.434174] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.909 [2024-11-26 19:05:32.446422] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.909 [2024-11-26 19:05:32.446437] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.909 [2024-11-26 19:05:32.458843] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.909 [2024-11-26 19:05:32.458858] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.909 [2024-11-26 19:05:32.471678] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.909 [2024-11-26 19:05:32.471693] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.909 [2024-11-26 19:05:32.484396] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.909 [2024-11-26 19:05:32.484411] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.909 [2024-11-26 19:05:32.497370] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.909 [2024-11-26 19:05:32.497385] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.909 [2024-11-26 19:05:32.510856] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.909 [2024-11-26 19:05:32.510877] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:19.909 [2024-11-26 19:05:32.524125] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:19.909 [2024-11-26 19:05:32.524141] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.536848] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.536869] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.549670] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.549689] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.562302] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.562317] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.575909] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.575924] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.588563] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.588578] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.601837] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.601852] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.614178] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.614193] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.627116] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.627131] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.640130] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.640145] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.653239] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.653254] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.666740] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.666755] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.679330] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.679345] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.691987] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.692002] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.705410] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.705425] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.718663] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.718679] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.732090] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.732106] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.745144] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.745159] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.758492] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.758508] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.771469] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.771485] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.172 [2024-11-26 19:05:32.784933] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.172 [2024-11-26 19:05:32.784948] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.798061] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.798080] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.811072] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.811087] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.824508] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.824524] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.837188] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.837204] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.849975] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.849991] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.862431] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.862446] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.875818] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.875834] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.888501] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.888517] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.901600] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.901616] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.915105] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.915119] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.927479] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.927494] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.940222] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.940237] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.952737] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.952753] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.966072] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.966088] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:32.979581] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.979596] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 19158.50 IOPS, 149.68 MiB/s [2024-11-26T18:05:33.060Z] [2024-11-26 19:05:32.992811] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:32.992827] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:33.006122] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:33.006137] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:33.018805] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:33.018821] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:33.031317] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:33.031333] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.435 [2024-11-26 19:05:33.044801] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.435 [2024-11-26 19:05:33.044816] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.057792] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.057808] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.070546] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.070562] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.084005] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.084021] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.097740] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.097756] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.110484] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.110499] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.123949] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.123968] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.137379] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.137395] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.150875] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.150891] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.164718] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.164734] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.177875] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.177890] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.190623] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.190639] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.203939] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.203955] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.216600] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.216615] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.229596] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.229612] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.242773] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.242789] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.255802] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.255818] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.268586] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.268602] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.281354] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.281369] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.294544] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.698 [2024-11-26 19:05:33.294559] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.698 [2024-11-26 19:05:33.308002] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.699 [2024-11-26 19:05:33.308017] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.321190] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.321205] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.334631] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.334647] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.347608] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.347623] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.361266] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.361282] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.374250] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.374265] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.386891] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.386906] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.399523] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.399538] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.413173] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.413189] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.426685] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.426700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.439178] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.439194] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.451885] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.451901] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.465473] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.465488] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.478531] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.478547] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.492289] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.492305] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.505203] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.505218] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.518025] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.518040] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.961 [2024-11-26 19:05:33.531525] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.961 [2024-11-26 19:05:33.531548] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.962 [2024-11-26 19:05:33.544837] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.962 [2024-11-26 19:05:33.544852] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.962 [2024-11-26 19:05:33.557878] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.962 [2024-11-26 19:05:33.557894] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:20.962 [2024-11-26 19:05:33.571224] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:20.962 [2024-11-26 19:05:33.571239] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.584147] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.584163] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.597641] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.597657] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.611105] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.611120] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.623746] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.623761] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.637020] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.637035] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.650230] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.650245] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.663629] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.663644] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.676728] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.676743] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.689852] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.689872] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.703384] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.703400] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.716955] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.716970] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.730310] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.730325] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.743640] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.743654] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.756328] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.756344] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.768677] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.768693] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.781954] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.781973] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.795334] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.795350] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.808217] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.808232] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.821763] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.821778] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.224 [2024-11-26 19:05:33.834173] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.224 [2024-11-26 19:05:33.834188] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.847110] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.847126] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.859966] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.859981] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.873357] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.873372] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.886852] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.886872] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.900174] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.900190] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.913087] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.913103] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.926439] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.926454] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.940065] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.940081] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.953459] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.953475] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.966109] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.966125] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:33.979045] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.979060] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 19173.00 IOPS, 149.79 MiB/s 00:09:21.486 Latency(us) 00:09:21.486 [2024-11-26T18:05:34.111Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:09:21.486 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:09:21.486 Nvme1n1 : 5.00 19183.23 149.87 0.00 0.00 6667.36 2635.09 16820.91 00:09:21.486 [2024-11-26T18:05:34.111Z] =================================================================================================================== 00:09:21.486 [2024-11-26T18:05:34.111Z] Total : 19183.23 149.87 0.00 0.00 6667.36 2635.09 16820.91 00:09:21.486 [2024-11-26 19:05:33.989109] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:33.989127] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.001137] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.001150] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.013174] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.013186] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.025202] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.025212] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.037230] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.037241] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.049257] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.049266] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.061287] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.061295] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.073319] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.073327] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.085352] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.085363] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 [2024-11-26 19:05:34.097381] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:09:21.486 [2024-11-26 19:05:34.097389] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:21.486 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (3692107) - No such process 00:09:21.486 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 3692107 00:09:21.486 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:21.486 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.486 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:21.748 delay0 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.748 19:05:34 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:09:21.748 [2024-11-26 19:05:34.299051] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:09:29.888 Initializing NVMe Controllers 00:09:29.888 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:09:29.888 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:09:29.888 Initialization complete. Launching workers. 00:09:29.888 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 215, failed: 38985 00:09:29.888 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 39045, failed to submit 155 00:09:29.888 success 38994, unsuccessful 51, failed 0 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@516 -- # nvmfcleanup 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:09:29.888 rmmod nvme_tcp 00:09:29.888 rmmod nvme_fabrics 00:09:29.888 rmmod nvme_keyring 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:09:29.888 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@517 -- # '[' -n 3689739 ']' 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@518 -- # killprocess 3689739 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@954 -- # '[' -z 3689739 ']' 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@958 -- # kill -0 3689739 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@959 -- # uname 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3689739 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3689739' 00:09:29.889 killing process with pid 3689739 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@973 -- # kill 3689739 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@978 -- # wait 3689739 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-save 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-restore 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:29.889 19:05:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:31.273 19:05:43 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:09:31.273 00:09:31.273 real 0m35.358s 00:09:31.273 user 0m45.987s 00:09:31.273 sys 0m12.200s 00:09:31.273 19:05:43 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:31.273 19:05:43 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:09:31.273 ************************************ 00:09:31.273 END TEST nvmf_zcopy 00:09:31.274 ************************************ 00:09:31.274 19:05:43 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:09:31.274 19:05:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:09:31.274 19:05:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:31.274 19:05:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:09:31.535 ************************************ 00:09:31.535 START TEST nvmf_nmic 00:09:31.535 ************************************ 00:09:31.535 19:05:43 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:09:31.535 * Looking for test storage... 00:09:31.535 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1693 -- # lcov --version 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:09:31.535 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:09:31.536 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:31.536 --rc genhtml_branch_coverage=1 00:09:31.536 --rc genhtml_function_coverage=1 00:09:31.536 --rc genhtml_legend=1 00:09:31.536 --rc geninfo_all_blocks=1 00:09:31.536 --rc geninfo_unexecuted_blocks=1 00:09:31.536 00:09:31.536 ' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:09:31.536 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:31.536 --rc genhtml_branch_coverage=1 00:09:31.536 --rc genhtml_function_coverage=1 00:09:31.536 --rc genhtml_legend=1 00:09:31.536 --rc geninfo_all_blocks=1 00:09:31.536 --rc geninfo_unexecuted_blocks=1 00:09:31.536 00:09:31.536 ' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:09:31.536 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:31.536 --rc genhtml_branch_coverage=1 00:09:31.536 --rc genhtml_function_coverage=1 00:09:31.536 --rc genhtml_legend=1 00:09:31.536 --rc geninfo_all_blocks=1 00:09:31.536 --rc geninfo_unexecuted_blocks=1 00:09:31.536 00:09:31.536 ' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:09:31.536 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:31.536 --rc genhtml_branch_coverage=1 00:09:31.536 --rc genhtml_function_coverage=1 00:09:31.536 --rc genhtml_legend=1 00:09:31.536 --rc geninfo_all_blocks=1 00:09:31.536 --rc geninfo_unexecuted_blocks=1 00:09:31.536 00:09:31.536 ' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:31.536 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@476 -- # prepare_net_devs 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@438 -- # local -g is_hw=no 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # remove_spdk_ns 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:09:31.536 19:05:44 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:09:39.683 Found 0000:31:00.0 (0x8086 - 0x159b) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:09:39.683 Found 0000:31:00.1 (0x8086 - 0x159b) 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:39.683 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:09:39.684 Found net devices under 0000:31:00.0: cvl_0_0 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:09:39.684 Found net devices under 0000:31:00.1: cvl_0_1 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # is_hw=yes 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:39.684 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:39.946 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:39.946 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:39.946 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:39.946 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:39.946 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:39.946 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:39.946 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:40.208 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:40.208 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:40.208 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.596 ms 00:09:40.208 00:09:40.208 --- 10.0.0.2 ping statistics --- 00:09:40.208 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:40.208 rtt min/avg/max/mdev = 0.596/0.596/0.596/0.000 ms 00:09:40.208 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:40.208 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:40.208 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.280 ms 00:09:40.208 00:09:40.208 --- 10.0.0.1 ping statistics --- 00:09:40.208 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:40.208 rtt min/avg/max/mdev = 0.280/0.280/0.280/0.000 ms 00:09:40.208 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:40.208 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@450 -- # return 0 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@509 -- # nvmfpid=3699477 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@510 -- # waitforlisten 3699477 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@835 -- # '[' -z 3699477 ']' 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:40.209 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:40.209 19:05:52 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:40.209 [2024-11-26 19:05:52.700348] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:09:40.209 [2024-11-26 19:05:52.700395] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:40.209 [2024-11-26 19:05:52.786480] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:40.209 [2024-11-26 19:05:52.823745] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:40.209 [2024-11-26 19:05:52.823778] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:40.209 [2024-11-26 19:05:52.823786] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:40.209 [2024-11-26 19:05:52.823793] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:40.209 [2024-11-26 19:05:52.823798] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:40.209 [2024-11-26 19:05:52.825273] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:40.209 [2024-11-26 19:05:52.825386] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:40.209 [2024-11-26 19:05:52.825539] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:40.209 [2024-11-26 19:05:52.825540] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@868 -- # return 0 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 [2024-11-26 19:05:53.552706] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 Malloc0 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 [2024-11-26 19:05:53.620359] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:09:41.207 test case1: single bdev can't be used in multiple subsystems 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.207 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.207 [2024-11-26 19:05:53.656273] bdev.c:8507:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:09:41.207 [2024-11-26 19:05:53.656295] subsystem.c:2156:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:09:41.207 [2024-11-26 19:05:53.656303] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:09:41.207 request: 00:09:41.207 { 00:09:41.207 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:09:41.207 "namespace": { 00:09:41.207 "bdev_name": "Malloc0", 00:09:41.207 "no_auto_visible": false, 00:09:41.207 "hide_metadata": false 00:09:41.207 }, 00:09:41.208 "method": "nvmf_subsystem_add_ns", 00:09:41.208 "req_id": 1 00:09:41.208 } 00:09:41.208 Got JSON-RPC error response 00:09:41.208 response: 00:09:41.208 { 00:09:41.208 "code": -32602, 00:09:41.208 "message": "Invalid parameters" 00:09:41.208 } 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:09:41.208 Adding namespace failed - expected result. 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:09:41.208 test case2: host connect to nvmf target in multiple paths 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:41.208 [2024-11-26 19:05:53.668421] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.208 19:05:53 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:09:42.705 19:05:55 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:09:44.087 19:05:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:09:44.087 19:05:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1202 -- # local i=0 00:09:44.087 19:05:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:09:44.087 19:05:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:09:44.087 19:05:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1209 -- # sleep 2 00:09:46.631 19:05:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:09:46.631 19:05:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:09:46.631 19:05:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:09:46.631 19:05:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:09:46.631 19:05:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:09:46.631 19:05:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1212 -- # return 0 00:09:46.631 19:05:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:09:46.631 [global] 00:09:46.631 thread=1 00:09:46.631 invalidate=1 00:09:46.631 rw=write 00:09:46.631 time_based=1 00:09:46.631 runtime=1 00:09:46.631 ioengine=libaio 00:09:46.631 direct=1 00:09:46.631 bs=4096 00:09:46.631 iodepth=1 00:09:46.631 norandommap=0 00:09:46.631 numjobs=1 00:09:46.631 00:09:46.631 verify_dump=1 00:09:46.631 verify_backlog=512 00:09:46.631 verify_state_save=0 00:09:46.631 do_verify=1 00:09:46.631 verify=crc32c-intel 00:09:46.631 [job0] 00:09:46.631 filename=/dev/nvme0n1 00:09:46.631 Could not set queue depth (nvme0n1) 00:09:46.631 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:46.631 fio-3.35 00:09:46.631 Starting 1 thread 00:09:48.018 00:09:48.018 job0: (groupid=0, jobs=1): err= 0: pid=3700798: Tue Nov 26 19:06:00 2024 00:09:48.018 read: IOPS=20, BW=82.7KiB/s (84.7kB/s)(84.0KiB/1016msec) 00:09:48.018 slat (nsec): min=7334, max=28622, avg=25135.62, stdev=5559.63 00:09:48.018 clat (usec): min=589, max=42134, avg=35942.29, stdev=14765.28 00:09:48.018 lat (usec): min=617, max=42161, avg=35967.43, stdev=14766.45 00:09:48.018 clat percentiles (usec): 00:09:48.018 | 1.00th=[ 586], 5.00th=[ 685], 10.00th=[ 693], 20.00th=[41157], 00:09:48.018 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[42206], 00:09:48.018 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:09:48.018 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:09:48.018 | 99.99th=[42206] 00:09:48.018 write: IOPS=503, BW=2016KiB/s (2064kB/s)(2048KiB/1016msec); 0 zone resets 00:09:48.018 slat (nsec): min=9114, max=71337, avg=30176.83, stdev=9984.29 00:09:48.018 clat (usec): min=155, max=1006, avg=472.04, stdev=113.05 00:09:48.018 lat (usec): min=185, max=1017, avg=502.21, stdev=117.99 00:09:48.018 clat percentiles (usec): 00:09:48.018 | 1.00th=[ 233], 5.00th=[ 269], 10.00th=[ 322], 20.00th=[ 375], 00:09:48.018 | 30.00th=[ 404], 40.00th=[ 457], 50.00th=[ 486], 60.00th=[ 506], 00:09:48.018 | 70.00th=[ 537], 80.00th=[ 570], 90.00th=[ 619], 95.00th=[ 644], 00:09:48.018 | 99.00th=[ 685], 99.50th=[ 693], 99.90th=[ 1004], 99.95th=[ 1004], 00:09:48.018 | 99.99th=[ 1004] 00:09:48.018 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:09:48.018 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:48.018 lat (usec) : 250=2.63%, 500=53.10%, 750=40.71% 00:09:48.018 lat (msec) : 2=0.19%, 50=3.38% 00:09:48.018 cpu : usr=0.89%, sys=2.07%, ctx=533, majf=0, minf=1 00:09:48.018 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:48.018 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:48.018 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:48.018 issued rwts: total=21,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:48.018 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:48.018 00:09:48.018 Run status group 0 (all jobs): 00:09:48.018 READ: bw=82.7KiB/s (84.7kB/s), 82.7KiB/s-82.7KiB/s (84.7kB/s-84.7kB/s), io=84.0KiB (86.0kB), run=1016-1016msec 00:09:48.018 WRITE: bw=2016KiB/s (2064kB/s), 2016KiB/s-2016KiB/s (2064kB/s-2064kB/s), io=2048KiB (2097kB), run=1016-1016msec 00:09:48.018 00:09:48.018 Disk stats (read/write): 00:09:48.018 nvme0n1: ios=68/512, merge=0/0, ticks=681/199, in_queue=880, util=93.29% 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:09:48.018 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1223 -- # local i=0 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1235 -- # return 0 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@516 -- # nvmfcleanup 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:09:48.018 rmmod nvme_tcp 00:09:48.018 rmmod nvme_fabrics 00:09:48.018 rmmod nvme_keyring 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@517 -- # '[' -n 3699477 ']' 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@518 -- # killprocess 3699477 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@954 -- # '[' -z 3699477 ']' 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@958 -- # kill -0 3699477 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@959 -- # uname 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3699477 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3699477' 00:09:48.018 killing process with pid 3699477 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@973 -- # kill 3699477 00:09:48.018 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@978 -- # wait 3699477 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-save 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-restore 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:48.280 19:06:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:50.192 19:06:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:09:50.192 00:09:50.192 real 0m18.889s 00:09:50.192 user 0m45.416s 00:09:50.192 sys 0m7.324s 00:09:50.192 19:06:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:50.192 19:06:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:50.192 ************************************ 00:09:50.192 END TEST nvmf_nmic 00:09:50.192 ************************************ 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:09:50.453 ************************************ 00:09:50.453 START TEST nvmf_fio_target 00:09:50.453 ************************************ 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:09:50.453 * Looking for test storage... 00:09:50.453 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1693 -- # lcov --version 00:09:50.453 19:06:02 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:09:50.453 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:09:50.713 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.713 --rc genhtml_branch_coverage=1 00:09:50.713 --rc genhtml_function_coverage=1 00:09:50.713 --rc genhtml_legend=1 00:09:50.713 --rc geninfo_all_blocks=1 00:09:50.713 --rc geninfo_unexecuted_blocks=1 00:09:50.713 00:09:50.713 ' 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:09:50.713 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.713 --rc genhtml_branch_coverage=1 00:09:50.713 --rc genhtml_function_coverage=1 00:09:50.713 --rc genhtml_legend=1 00:09:50.713 --rc geninfo_all_blocks=1 00:09:50.713 --rc geninfo_unexecuted_blocks=1 00:09:50.713 00:09:50.713 ' 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:09:50.713 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.713 --rc genhtml_branch_coverage=1 00:09:50.713 --rc genhtml_function_coverage=1 00:09:50.713 --rc genhtml_legend=1 00:09:50.713 --rc geninfo_all_blocks=1 00:09:50.713 --rc geninfo_unexecuted_blocks=1 00:09:50.713 00:09:50.713 ' 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:09:50.713 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.713 --rc genhtml_branch_coverage=1 00:09:50.713 --rc genhtml_function_coverage=1 00:09:50.713 --rc genhtml_legend=1 00:09:50.713 --rc geninfo_all_blocks=1 00:09:50.713 --rc geninfo_unexecuted_blocks=1 00:09:50.713 00:09:50.713 ' 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:50.713 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:50.714 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:09:50.714 19:06:03 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:09:58.853 Found 0000:31:00.0 (0x8086 - 0x159b) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:09:58.853 Found 0000:31:00.1 (0x8086 - 0x159b) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:09:58.853 Found net devices under 0000:31:00.0: cvl_0_0 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:09:58.853 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:09:58.854 Found net devices under 0000:31:00.1: cvl_0_1 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # is_hw=yes 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:58.854 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:59.114 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:59.114 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.614 ms 00:09:59.114 00:09:59.114 --- 10.0.0.2 ping statistics --- 00:09:59.114 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:59.114 rtt min/avg/max/mdev = 0.614/0.614/0.614/0.000 ms 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:59.114 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:59.114 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.286 ms 00:09:59.114 00:09:59.114 --- 10.0.0.1 ping statistics --- 00:09:59.114 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:59.114 rtt min/avg/max/mdev = 0.286/0.286/0.286/0.000 ms 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@450 -- # return 0 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@509 -- # nvmfpid=3706614 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@510 -- # waitforlisten 3706614 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@835 -- # '[' -z 3706614 ']' 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:59.114 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:59.114 19:06:11 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:59.114 [2024-11-26 19:06:11.716514] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:09:59.114 [2024-11-26 19:06:11.716566] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:59.378 [2024-11-26 19:06:11.802415] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:59.378 [2024-11-26 19:06:11.837957] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:59.378 [2024-11-26 19:06:11.837989] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:59.378 [2024-11-26 19:06:11.837997] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:59.378 [2024-11-26 19:06:11.838003] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:59.378 [2024-11-26 19:06:11.838009] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:59.378 [2024-11-26 19:06:11.839509] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:59.378 [2024-11-26 19:06:11.839629] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:59.378 [2024-11-26 19:06:11.839783] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:59.378 [2024-11-26 19:06:11.839783] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:59.948 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:59.948 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@868 -- # return 0 00:09:59.948 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:09:59.949 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:59.949 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:59.949 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:59.949 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:10:00.209 [2024-11-26 19:06:12.707108] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:00.209 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:00.469 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:10:00.469 19:06:12 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:00.729 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:10:00.729 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:00.729 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:10:00.729 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:00.990 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:10:00.990 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:10:01.251 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:01.511 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:10:01.511 19:06:13 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:01.511 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:10:01.511 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:01.771 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:10:01.771 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:10:02.031 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:10:02.031 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:10:02.031 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:10:02.292 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:10:02.292 19:06:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:10:02.552 19:06:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:02.552 [2024-11-26 19:06:15.160481] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:02.813 19:06:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:10:02.813 19:06:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:10:03.073 19:06:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:10:04.453 19:06:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:10:04.453 19:06:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1202 -- # local i=0 00:10:04.453 19:06:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:10:04.453 19:06:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1204 -- # [[ -n 4 ]] 00:10:04.453 19:06:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1205 -- # nvme_device_counter=4 00:10:04.453 19:06:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1209 -- # sleep 2 00:10:06.993 19:06:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:10:06.993 19:06:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:10:06.993 19:06:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:10:06.993 19:06:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1211 -- # nvme_devices=4 00:10:06.993 19:06:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:10:06.993 19:06:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1212 -- # return 0 00:10:06.993 19:06:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:10:06.993 [global] 00:10:06.993 thread=1 00:10:06.993 invalidate=1 00:10:06.993 rw=write 00:10:06.993 time_based=1 00:10:06.993 runtime=1 00:10:06.993 ioengine=libaio 00:10:06.993 direct=1 00:10:06.993 bs=4096 00:10:06.993 iodepth=1 00:10:06.993 norandommap=0 00:10:06.993 numjobs=1 00:10:06.993 00:10:06.993 verify_dump=1 00:10:06.993 verify_backlog=512 00:10:06.993 verify_state_save=0 00:10:06.993 do_verify=1 00:10:06.993 verify=crc32c-intel 00:10:06.993 [job0] 00:10:06.993 filename=/dev/nvme0n1 00:10:06.993 [job1] 00:10:06.993 filename=/dev/nvme0n2 00:10:06.993 [job2] 00:10:06.993 filename=/dev/nvme0n3 00:10:06.993 [job3] 00:10:06.993 filename=/dev/nvme0n4 00:10:06.993 Could not set queue depth (nvme0n1) 00:10:06.993 Could not set queue depth (nvme0n2) 00:10:06.993 Could not set queue depth (nvme0n3) 00:10:06.993 Could not set queue depth (nvme0n4) 00:10:06.993 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:06.993 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:06.993 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:06.993 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:06.993 fio-3.35 00:10:06.993 Starting 4 threads 00:10:08.375 00:10:08.375 job0: (groupid=0, jobs=1): err= 0: pid=3708244: Tue Nov 26 19:06:20 2024 00:10:08.375 read: IOPS=56, BW=225KiB/s (231kB/s)(228KiB/1012msec) 00:10:08.375 slat (nsec): min=11061, max=52981, avg=29145.95, stdev=4759.75 00:10:08.375 clat (usec): min=429, max=42312, avg=11489.94, stdev=18441.50 00:10:08.375 lat (usec): min=460, max=42340, avg=11519.08, stdev=18441.75 00:10:08.375 clat percentiles (usec): 00:10:08.375 | 1.00th=[ 429], 5.00th=[ 457], 10.00th=[ 478], 20.00th=[ 498], 00:10:08.375 | 30.00th=[ 578], 40.00th=[ 586], 50.00th=[ 611], 60.00th=[ 627], 00:10:08.375 | 70.00th=[ 668], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:10:08.375 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:10:08.375 | 99.99th=[42206] 00:10:08.375 write: IOPS=505, BW=2024KiB/s (2072kB/s)(2048KiB/1012msec); 0 zone resets 00:10:08.375 slat (nsec): min=9782, max=60913, avg=31270.94, stdev=11077.62 00:10:08.375 clat (usec): min=116, max=1277, avg=648.32, stdev=159.59 00:10:08.375 lat (usec): min=129, max=1324, avg=679.59, stdev=165.04 00:10:08.375 clat percentiles (usec): 00:10:08.375 | 1.00th=[ 210], 5.00th=[ 306], 10.00th=[ 420], 20.00th=[ 545], 00:10:08.375 | 30.00th=[ 611], 40.00th=[ 644], 50.00th=[ 676], 60.00th=[ 709], 00:10:08.375 | 70.00th=[ 734], 80.00th=[ 775], 90.00th=[ 816], 95.00th=[ 857], 00:10:08.375 | 99.00th=[ 930], 99.50th=[ 955], 99.90th=[ 1270], 99.95th=[ 1270], 00:10:08.375 | 99.99th=[ 1270] 00:10:08.375 bw ( KiB/s): min= 4096, max= 4096, per=40.48%, avg=4096.00, stdev= 0.00, samples=1 00:10:08.375 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:08.375 lat (usec) : 250=2.11%, 500=13.88%, 750=58.00%, 1000=23.02% 00:10:08.375 lat (msec) : 2=0.35%, 50=2.64% 00:10:08.375 cpu : usr=1.29%, sys=1.88%, ctx=573, majf=0, minf=1 00:10:08.375 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:08.375 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.375 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.375 issued rwts: total=57,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:08.375 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:08.375 job1: (groupid=0, jobs=1): err= 0: pid=3708245: Tue Nov 26 19:06:20 2024 00:10:08.375 read: IOPS=15, BW=63.9KiB/s (65.4kB/s)(64.0KiB/1002msec) 00:10:08.375 slat (nsec): min=25498, max=26869, avg=25771.31, stdev=349.82 00:10:08.375 clat (usec): min=41874, max=44002, avg=42091.41, stdev=511.47 00:10:08.375 lat (usec): min=41899, max=44029, avg=42117.18, stdev=511.76 00:10:08.375 clat percentiles (usec): 00:10:08.375 | 1.00th=[41681], 5.00th=[41681], 10.00th=[41681], 20.00th=[42206], 00:10:08.375 | 30.00th=[42206], 40.00th=[42206], 50.00th=[42206], 60.00th=[42206], 00:10:08.375 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[43779], 00:10:08.375 | 99.00th=[43779], 99.50th=[43779], 99.90th=[43779], 99.95th=[43779], 00:10:08.376 | 99.99th=[43779] 00:10:08.376 write: IOPS=510, BW=2044KiB/s (2093kB/s)(2048KiB/1002msec); 0 zone resets 00:10:08.376 slat (nsec): min=9885, max=54607, avg=30885.04, stdev=9273.94 00:10:08.376 clat (usec): min=137, max=943, avg=602.03, stdev=113.66 00:10:08.376 lat (usec): min=149, max=977, avg=632.92, stdev=117.42 00:10:08.376 clat percentiles (usec): 00:10:08.376 | 1.00th=[ 343], 5.00th=[ 388], 10.00th=[ 457], 20.00th=[ 506], 00:10:08.376 | 30.00th=[ 553], 40.00th=[ 578], 50.00th=[ 611], 60.00th=[ 635], 00:10:08.376 | 70.00th=[ 668], 80.00th=[ 701], 90.00th=[ 742], 95.00th=[ 775], 00:10:08.376 | 99.00th=[ 816], 99.50th=[ 832], 99.90th=[ 947], 99.95th=[ 947], 00:10:08.376 | 99.99th=[ 947] 00:10:08.376 bw ( KiB/s): min= 4096, max= 4096, per=40.48%, avg=4096.00, stdev= 0.00, samples=1 00:10:08.376 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:08.376 lat (usec) : 250=0.19%, 500=18.56%, 750=70.27%, 1000=7.95% 00:10:08.376 lat (msec) : 50=3.03% 00:10:08.376 cpu : usr=1.00%, sys=1.20%, ctx=531, majf=0, minf=1 00:10:08.376 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:08.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.376 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.376 issued rwts: total=16,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:08.376 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:08.376 job2: (groupid=0, jobs=1): err= 0: pid=3708264: Tue Nov 26 19:06:20 2024 00:10:08.376 read: IOPS=821, BW=3285KiB/s (3364kB/s)(3288KiB/1001msec) 00:10:08.376 slat (nsec): min=7007, max=51355, avg=23602.15, stdev=8715.38 00:10:08.376 clat (usec): min=190, max=42209, avg=772.33, stdev=2836.30 00:10:08.376 lat (usec): min=200, max=42236, avg=795.94, stdev=2836.65 00:10:08.376 clat percentiles (usec): 00:10:08.376 | 1.00th=[ 310], 5.00th=[ 412], 10.00th=[ 469], 20.00th=[ 502], 00:10:08.376 | 30.00th=[ 537], 40.00th=[ 578], 50.00th=[ 594], 60.00th=[ 611], 00:10:08.376 | 70.00th=[ 627], 80.00th=[ 644], 90.00th=[ 668], 95.00th=[ 685], 00:10:08.376 | 99.00th=[ 742], 99.50th=[ 2008], 99.90th=[42206], 99.95th=[42206], 00:10:08.376 | 99.99th=[42206] 00:10:08.376 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:10:08.376 slat (nsec): min=10249, max=70835, avg=27436.03, stdev=11779.32 00:10:08.376 clat (usec): min=98, max=665, avg=295.94, stdev=67.96 00:10:08.376 lat (usec): min=109, max=676, avg=323.37, stdev=69.72 00:10:08.376 clat percentiles (usec): 00:10:08.376 | 1.00th=[ 111], 5.00th=[ 192], 10.00th=[ 208], 20.00th=[ 243], 00:10:08.376 | 30.00th=[ 273], 40.00th=[ 293], 50.00th=[ 302], 60.00th=[ 310], 00:10:08.376 | 70.00th=[ 330], 80.00th=[ 351], 90.00th=[ 375], 95.00th=[ 400], 00:10:08.376 | 99.00th=[ 445], 99.50th=[ 457], 99.90th=[ 474], 99.95th=[ 668], 00:10:08.376 | 99.99th=[ 668] 00:10:08.376 bw ( KiB/s): min= 4096, max= 4096, per=40.48%, avg=4096.00, stdev= 0.00, samples=1 00:10:08.376 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:08.376 lat (usec) : 100=0.05%, 250=12.57%, 500=51.52%, 750=35.48%, 1000=0.11% 00:10:08.376 lat (msec) : 4=0.05%, 50=0.22% 00:10:08.376 cpu : usr=2.40%, sys=4.90%, ctx=1848, majf=0, minf=1 00:10:08.376 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:08.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.376 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.376 issued rwts: total=822,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:08.376 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:08.376 job3: (groupid=0, jobs=1): err= 0: pid=3708271: Tue Nov 26 19:06:20 2024 00:10:08.376 read: IOPS=80, BW=320KiB/s (328kB/s)(324KiB/1011msec) 00:10:08.376 slat (nsec): min=7989, max=51032, avg=26485.70, stdev=7641.77 00:10:08.376 clat (usec): min=426, max=42308, avg=10012.72, stdev=17304.92 00:10:08.376 lat (usec): min=454, max=42340, avg=10039.21, stdev=17307.64 00:10:08.376 clat percentiles (usec): 00:10:08.376 | 1.00th=[ 429], 5.00th=[ 461], 10.00th=[ 482], 20.00th=[ 510], 00:10:08.376 | 30.00th=[ 545], 40.00th=[ 586], 50.00th=[ 603], 60.00th=[ 627], 00:10:08.376 | 70.00th=[ 668], 80.00th=[40633], 90.00th=[42206], 95.00th=[42206], 00:10:08.376 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:10:08.376 | 99.99th=[42206] 00:10:08.376 write: IOPS=506, BW=2026KiB/s (2074kB/s)(2048KiB/1011msec); 0 zone resets 00:10:08.376 slat (nsec): min=10583, max=58504, avg=30302.88, stdev=11315.84 00:10:08.376 clat (usec): min=108, max=667, avg=342.68, stdev=90.51 00:10:08.376 lat (usec): min=119, max=683, avg=372.99, stdev=93.84 00:10:08.376 clat percentiles (usec): 00:10:08.376 | 1.00th=[ 119], 5.00th=[ 155], 10.00th=[ 245], 20.00th=[ 273], 00:10:08.376 | 30.00th=[ 293], 40.00th=[ 322], 50.00th=[ 355], 60.00th=[ 375], 00:10:08.376 | 70.00th=[ 396], 80.00th=[ 420], 90.00th=[ 449], 95.00th=[ 465], 00:10:08.376 | 99.00th=[ 529], 99.50th=[ 562], 99.90th=[ 668], 99.95th=[ 668], 00:10:08.376 | 99.99th=[ 668] 00:10:08.376 bw ( KiB/s): min= 4096, max= 4096, per=40.48%, avg=4096.00, stdev= 0.00, samples=1 00:10:08.376 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:08.376 lat (usec) : 250=9.95%, 500=76.22%, 750=10.29%, 1000=0.34% 00:10:08.376 lat (msec) : 50=3.20% 00:10:08.376 cpu : usr=1.09%, sys=1.49%, ctx=594, majf=0, minf=1 00:10:08.376 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:08.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.376 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:08.376 issued rwts: total=81,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:08.376 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:08.376 00:10:08.376 Run status group 0 (all jobs): 00:10:08.376 READ: bw=3858KiB/s (3950kB/s), 63.9KiB/s-3285KiB/s (65.4kB/s-3364kB/s), io=3904KiB (3998kB), run=1001-1012msec 00:10:08.376 WRITE: bw=9.88MiB/s (10.4MB/s), 2024KiB/s-4092KiB/s (2072kB/s-4190kB/s), io=10.0MiB (10.5MB), run=1001-1012msec 00:10:08.376 00:10:08.376 Disk stats (read/write): 00:10:08.376 nvme0n1: ios=62/512, merge=0/0, ticks=1268/278, in_queue=1546, util=83.67% 00:10:08.376 nvme0n2: ios=33/512, merge=0/0, ticks=1335/299, in_queue=1634, util=87.83% 00:10:08.376 nvme0n3: ios=567/1024, merge=0/0, ticks=1328/295, in_queue=1623, util=91.74% 00:10:08.376 nvme0n4: ios=127/512, merge=0/0, ticks=727/167, in_queue=894, util=97.32% 00:10:08.376 19:06:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:10:08.376 [global] 00:10:08.376 thread=1 00:10:08.376 invalidate=1 00:10:08.376 rw=randwrite 00:10:08.376 time_based=1 00:10:08.376 runtime=1 00:10:08.376 ioengine=libaio 00:10:08.376 direct=1 00:10:08.376 bs=4096 00:10:08.376 iodepth=1 00:10:08.376 norandommap=0 00:10:08.376 numjobs=1 00:10:08.376 00:10:08.376 verify_dump=1 00:10:08.376 verify_backlog=512 00:10:08.376 verify_state_save=0 00:10:08.376 do_verify=1 00:10:08.376 verify=crc32c-intel 00:10:08.376 [job0] 00:10:08.376 filename=/dev/nvme0n1 00:10:08.376 [job1] 00:10:08.376 filename=/dev/nvme0n2 00:10:08.376 [job2] 00:10:08.376 filename=/dev/nvme0n3 00:10:08.376 [job3] 00:10:08.376 filename=/dev/nvme0n4 00:10:08.376 Could not set queue depth (nvme0n1) 00:10:08.376 Could not set queue depth (nvme0n2) 00:10:08.376 Could not set queue depth (nvme0n3) 00:10:08.376 Could not set queue depth (nvme0n4) 00:10:08.638 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:08.638 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:08.638 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:08.638 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:08.638 fio-3.35 00:10:08.638 Starting 4 threads 00:10:10.021 00:10:10.021 job0: (groupid=0, jobs=1): err= 0: pid=3708769: Tue Nov 26 19:06:22 2024 00:10:10.021 read: IOPS=28, BW=116KiB/s (119kB/s)(116KiB/1001msec) 00:10:10.021 slat (nsec): min=26091, max=48343, avg=27376.66, stdev=4041.14 00:10:10.021 clat (usec): min=781, max=42970, avg=23505.49, stdev=20652.45 00:10:10.021 lat (usec): min=808, max=42997, avg=23532.87, stdev=20651.61 00:10:10.021 clat percentiles (usec): 00:10:10.021 | 1.00th=[ 783], 5.00th=[ 906], 10.00th=[ 938], 20.00th=[ 1004], 00:10:10.021 | 30.00th=[ 1037], 40.00th=[ 1090], 50.00th=[41157], 60.00th=[41157], 00:10:10.021 | 70.00th=[41681], 80.00th=[42206], 90.00th=[42730], 95.00th=[42730], 00:10:10.021 | 99.00th=[42730], 99.50th=[42730], 99.90th=[42730], 99.95th=[42730], 00:10:10.021 | 99.99th=[42730] 00:10:10.021 write: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec); 0 zone resets 00:10:10.021 slat (nsec): min=9696, max=72117, avg=30806.60, stdev=9870.37 00:10:10.021 clat (usec): min=275, max=930, avg=582.11, stdev=118.08 00:10:10.021 lat (usec): min=285, max=977, avg=612.92, stdev=122.03 00:10:10.021 clat percentiles (usec): 00:10:10.021 | 1.00th=[ 314], 5.00th=[ 379], 10.00th=[ 416], 20.00th=[ 486], 00:10:10.021 | 30.00th=[ 529], 40.00th=[ 553], 50.00th=[ 586], 60.00th=[ 611], 00:10:10.021 | 70.00th=[ 644], 80.00th=[ 685], 90.00th=[ 725], 95.00th=[ 766], 00:10:10.021 | 99.00th=[ 873], 99.50th=[ 898], 99.90th=[ 930], 99.95th=[ 930], 00:10:10.021 | 99.99th=[ 930] 00:10:10.021 bw ( KiB/s): min= 4096, max= 4096, per=36.88%, avg=4096.00, stdev= 0.00, samples=1 00:10:10.021 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:10.021 lat (usec) : 500=22.37%, 750=66.36%, 1000=6.84% 00:10:10.021 lat (msec) : 2=1.48%, 50=2.96% 00:10:10.021 cpu : usr=1.10%, sys=1.30%, ctx=543, majf=0, minf=1 00:10:10.021 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:10.021 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.021 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.021 issued rwts: total=29,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:10.021 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:10.021 job1: (groupid=0, jobs=1): err= 0: pid=3708775: Tue Nov 26 19:06:22 2024 00:10:10.021 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:10:10.021 slat (nsec): min=25475, max=45136, avg=26546.80, stdev=2415.72 00:10:10.021 clat (usec): min=725, max=1431, avg=1118.03, stdev=115.05 00:10:10.021 lat (usec): min=751, max=1458, avg=1144.58, stdev=114.89 00:10:10.021 clat percentiles (usec): 00:10:10.021 | 1.00th=[ 775], 5.00th=[ 914], 10.00th=[ 963], 20.00th=[ 1029], 00:10:10.021 | 30.00th=[ 1074], 40.00th=[ 1106], 50.00th=[ 1123], 60.00th=[ 1156], 00:10:10.021 | 70.00th=[ 1188], 80.00th=[ 1221], 90.00th=[ 1254], 95.00th=[ 1287], 00:10:10.021 | 99.00th=[ 1336], 99.50th=[ 1385], 99.90th=[ 1434], 99.95th=[ 1434], 00:10:10.021 | 99.99th=[ 1434] 00:10:10.021 write: IOPS=661, BW=2645KiB/s (2709kB/s)(2648KiB/1001msec); 0 zone resets 00:10:10.021 slat (nsec): min=9766, max=97303, avg=32289.48, stdev=7607.47 00:10:10.021 clat (usec): min=154, max=1010, avg=577.72, stdev=142.43 00:10:10.021 lat (usec): min=164, max=1023, avg=610.01, stdev=143.82 00:10:10.021 clat percentiles (usec): 00:10:10.021 | 1.00th=[ 269], 5.00th=[ 347], 10.00th=[ 388], 20.00th=[ 457], 00:10:10.021 | 30.00th=[ 494], 40.00th=[ 537], 50.00th=[ 578], 60.00th=[ 619], 00:10:10.021 | 70.00th=[ 652], 80.00th=[ 693], 90.00th=[ 766], 95.00th=[ 807], 00:10:10.021 | 99.00th=[ 906], 99.50th=[ 938], 99.90th=[ 1012], 99.95th=[ 1012], 00:10:10.021 | 99.99th=[ 1012] 00:10:10.021 bw ( KiB/s): min= 4096, max= 4096, per=36.88%, avg=4096.00, stdev= 0.00, samples=1 00:10:10.021 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:10.021 lat (usec) : 250=0.34%, 500=17.29%, 750=32.45%, 1000=12.78% 00:10:10.021 lat (msec) : 2=37.14% 00:10:10.021 cpu : usr=2.10%, sys=3.40%, ctx=1177, majf=0, minf=1 00:10:10.021 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:10.021 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.021 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.021 issued rwts: total=512,662,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:10.022 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:10.022 job2: (groupid=0, jobs=1): err= 0: pid=3708789: Tue Nov 26 19:06:22 2024 00:10:10.022 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:10:10.022 slat (nsec): min=7071, max=61255, avg=27168.27, stdev=5102.47 00:10:10.022 clat (usec): min=376, max=42002, avg=1176.44, stdev=4248.19 00:10:10.022 lat (usec): min=403, max=42029, avg=1203.60, stdev=4248.21 00:10:10.022 clat percentiles (usec): 00:10:10.022 | 1.00th=[ 404], 5.00th=[ 445], 10.00th=[ 515], 20.00th=[ 586], 00:10:10.022 | 30.00th=[ 611], 40.00th=[ 644], 50.00th=[ 742], 60.00th=[ 783], 00:10:10.022 | 70.00th=[ 840], 80.00th=[ 873], 90.00th=[ 906], 95.00th=[ 930], 00:10:10.022 | 99.00th=[33162], 99.50th=[41157], 99.90th=[42206], 99.95th=[42206], 00:10:10.022 | 99.99th=[42206] 00:10:10.022 write: IOPS=658, BW=2633KiB/s (2697kB/s)(2636KiB/1001msec); 0 zone resets 00:10:10.022 slat (usec): min=3, max=25567, avg=66.98, stdev=994.94 00:10:10.022 clat (usec): min=184, max=2485, avg=502.07, stdev=155.19 00:10:10.022 lat (usec): min=218, max=26367, avg=569.04, stdev=1018.33 00:10:10.022 clat percentiles (usec): 00:10:10.022 | 1.00th=[ 265], 5.00th=[ 310], 10.00th=[ 338], 20.00th=[ 379], 00:10:10.022 | 30.00th=[ 420], 40.00th=[ 457], 50.00th=[ 490], 60.00th=[ 519], 00:10:10.022 | 70.00th=[ 553], 80.00th=[ 603], 90.00th=[ 676], 95.00th=[ 766], 00:10:10.022 | 99.00th=[ 938], 99.50th=[ 963], 99.90th=[ 2474], 99.95th=[ 2474], 00:10:10.022 | 99.99th=[ 2474] 00:10:10.022 bw ( KiB/s): min= 4096, max= 4096, per=36.88%, avg=4096.00, stdev= 0.00, samples=1 00:10:10.022 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:10.022 lat (usec) : 250=0.34%, 500=34.16%, 750=42.27%, 1000=22.46% 00:10:10.022 lat (msec) : 2=0.17%, 4=0.09%, 50=0.51% 00:10:10.022 cpu : usr=1.50%, sys=3.90%, ctx=1176, majf=0, minf=1 00:10:10.022 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:10.022 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.022 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.022 issued rwts: total=512,659,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:10.022 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:10.022 job3: (groupid=0, jobs=1): err= 0: pid=3708796: Tue Nov 26 19:06:22 2024 00:10:10.022 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:10:10.022 slat (nsec): min=6577, max=47602, avg=25344.29, stdev=8519.33 00:10:10.022 clat (usec): min=685, max=1202, avg=958.23, stdev=85.12 00:10:10.022 lat (usec): min=712, max=1228, avg=983.57, stdev=88.70 00:10:10.022 clat percentiles (usec): 00:10:10.022 | 1.00th=[ 717], 5.00th=[ 799], 10.00th=[ 840], 20.00th=[ 889], 00:10:10.022 | 30.00th=[ 922], 40.00th=[ 955], 50.00th=[ 971], 60.00th=[ 988], 00:10:10.022 | 70.00th=[ 1004], 80.00th=[ 1020], 90.00th=[ 1057], 95.00th=[ 1074], 00:10:10.022 | 99.00th=[ 1139], 99.50th=[ 1156], 99.90th=[ 1205], 99.95th=[ 1205], 00:10:10.022 | 99.99th=[ 1205] 00:10:10.022 write: IOPS=945, BW=3780KiB/s (3871kB/s)(3784KiB/1001msec); 0 zone resets 00:10:10.022 slat (nsec): min=9330, max=53763, avg=29858.66, stdev=10939.51 00:10:10.022 clat (usec): min=196, max=795, avg=483.59, stdev=121.44 00:10:10.022 lat (usec): min=215, max=831, avg=513.45, stdev=123.69 00:10:10.022 clat percentiles (usec): 00:10:10.022 | 1.00th=[ 217], 5.00th=[ 302], 10.00th=[ 330], 20.00th=[ 359], 00:10:10.022 | 30.00th=[ 420], 40.00th=[ 453], 50.00th=[ 482], 60.00th=[ 510], 00:10:10.022 | 70.00th=[ 553], 80.00th=[ 594], 90.00th=[ 644], 95.00th=[ 685], 00:10:10.022 | 99.00th=[ 758], 99.50th=[ 766], 99.90th=[ 799], 99.95th=[ 799], 00:10:10.022 | 99.99th=[ 799] 00:10:10.022 bw ( KiB/s): min= 4096, max= 4096, per=36.88%, avg=4096.00, stdev= 0.00, samples=1 00:10:10.022 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:10:10.022 lat (usec) : 250=1.85%, 500=34.71%, 750=28.19%, 1000=24.01% 00:10:10.022 lat (msec) : 2=11.25% 00:10:10.022 cpu : usr=3.00%, sys=4.60%, ctx=1460, majf=0, minf=1 00:10:10.022 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:10.022 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.022 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:10.022 issued rwts: total=512,946,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:10.022 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:10.022 00:10:10.022 Run status group 0 (all jobs): 00:10:10.022 READ: bw=6254KiB/s (6404kB/s), 116KiB/s-2046KiB/s (119kB/s-2095kB/s), io=6260KiB (6410kB), run=1001-1001msec 00:10:10.022 WRITE: bw=10.8MiB/s (11.4MB/s), 2046KiB/s-3780KiB/s (2095kB/s-3871kB/s), io=10.9MiB (11.4MB), run=1001-1001msec 00:10:10.022 00:10:10.022 Disk stats (read/write): 00:10:10.022 nvme0n1: ios=38/512, merge=0/0, ticks=1341/275, in_queue=1616, util=84.27% 00:10:10.022 nvme0n2: ios=512/512, merge=0/0, ticks=598/265, in_queue=863, util=91.23% 00:10:10.022 nvme0n3: ios=443/512, merge=0/0, ticks=750/240, in_queue=990, util=94.83% 00:10:10.022 nvme0n4: ios=573/681, merge=0/0, ticks=701/289, in_queue=990, util=94.76% 00:10:10.022 19:06:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:10:10.022 [global] 00:10:10.022 thread=1 00:10:10.022 invalidate=1 00:10:10.022 rw=write 00:10:10.022 time_based=1 00:10:10.022 runtime=1 00:10:10.022 ioengine=libaio 00:10:10.022 direct=1 00:10:10.022 bs=4096 00:10:10.022 iodepth=128 00:10:10.022 norandommap=0 00:10:10.022 numjobs=1 00:10:10.022 00:10:10.022 verify_dump=1 00:10:10.022 verify_backlog=512 00:10:10.022 verify_state_save=0 00:10:10.022 do_verify=1 00:10:10.022 verify=crc32c-intel 00:10:10.022 [job0] 00:10:10.022 filename=/dev/nvme0n1 00:10:10.022 [job1] 00:10:10.022 filename=/dev/nvme0n2 00:10:10.022 [job2] 00:10:10.022 filename=/dev/nvme0n3 00:10:10.022 [job3] 00:10:10.022 filename=/dev/nvme0n4 00:10:10.022 Could not set queue depth (nvme0n1) 00:10:10.022 Could not set queue depth (nvme0n2) 00:10:10.022 Could not set queue depth (nvme0n3) 00:10:10.022 Could not set queue depth (nvme0n4) 00:10:10.283 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:10.283 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:10.283 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:10.283 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:10.283 fio-3.35 00:10:10.283 Starting 4 threads 00:10:11.668 00:10:11.668 job0: (groupid=0, jobs=1): err= 0: pid=3709273: Tue Nov 26 19:06:24 2024 00:10:11.668 read: IOPS=7664, BW=29.9MiB/s (31.4MB/s)(30.0MiB/1002msec) 00:10:11.668 slat (nsec): min=898, max=7495.1k, avg=63730.93, stdev=450633.94 00:10:11.668 clat (usec): min=3391, max=22849, avg=8959.16, stdev=3096.75 00:10:11.668 lat (usec): min=3400, max=22855, avg=9022.89, stdev=3126.14 00:10:11.668 clat percentiles (usec): 00:10:11.668 | 1.00th=[ 3621], 5.00th=[ 5604], 10.00th=[ 6456], 20.00th=[ 6915], 00:10:11.668 | 30.00th=[ 7308], 40.00th=[ 7832], 50.00th=[ 8356], 60.00th=[ 8717], 00:10:11.668 | 70.00th=[ 9241], 80.00th=[10290], 90.00th=[12125], 95.00th=[14877], 00:10:11.668 | 99.00th=[21627], 99.50th=[22414], 99.90th=[22938], 99.95th=[22938], 00:10:11.668 | 99.99th=[22938] 00:10:11.668 write: IOPS=7734, BW=30.2MiB/s (31.7MB/s)(30.3MiB/1002msec); 0 zone resets 00:10:11.668 slat (nsec): min=1599, max=6732.4k, avg=50508.67, stdev=383000.17 00:10:11.668 clat (usec): min=433, max=29732, avg=7528.03, stdev=2853.50 00:10:11.668 lat (usec): min=1019, max=29734, avg=7578.54, stdev=2872.76 00:10:11.668 clat percentiles (usec): 00:10:11.668 | 1.00th=[ 2212], 5.00th=[ 3556], 10.00th=[ 4424], 20.00th=[ 5800], 00:10:11.668 | 30.00th=[ 6521], 40.00th=[ 6849], 50.00th=[ 7177], 60.00th=[ 7439], 00:10:11.668 | 70.00th=[ 8029], 80.00th=[ 8717], 90.00th=[10814], 95.00th=[12256], 00:10:11.668 | 99.00th=[18744], 99.50th=[21103], 99.90th=[22938], 99.95th=[22938], 00:10:11.668 | 99.99th=[29754] 00:10:11.668 bw ( KiB/s): min=28672, max=32768, per=31.58%, avg=30720.00, stdev=2896.31, samples=2 00:10:11.668 iops : min= 7168, max= 8192, avg=7680.00, stdev=724.08, samples=2 00:10:11.668 lat (usec) : 500=0.01% 00:10:11.668 lat (msec) : 2=0.42%, 4=4.26%, 10=76.82%, 20=16.95%, 50=1.54% 00:10:11.668 cpu : usr=5.19%, sys=9.19%, ctx=435, majf=0, minf=2 00:10:11.668 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:10:11.668 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:11.668 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:11.668 issued rwts: total=7680,7750,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:11.668 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:11.668 job1: (groupid=0, jobs=1): err= 0: pid=3709288: Tue Nov 26 19:06:24 2024 00:10:11.668 read: IOPS=5557, BW=21.7MiB/s (22.8MB/s)(21.9MiB/1010msec) 00:10:11.668 slat (nsec): min=945, max=11983k, avg=99183.88, stdev=714292.57 00:10:11.668 clat (usec): min=3746, max=48061, avg=12041.66, stdev=4055.83 00:10:11.668 lat (usec): min=3751, max=48069, avg=12140.84, stdev=4117.92 00:10:11.668 clat percentiles (usec): 00:10:11.668 | 1.00th=[ 4293], 5.00th=[ 8586], 10.00th=[ 9634], 20.00th=[ 9896], 00:10:11.668 | 30.00th=[10159], 40.00th=[10683], 50.00th=[10945], 60.00th=[11600], 00:10:11.668 | 70.00th=[12780], 80.00th=[13435], 90.00th=[15795], 95.00th=[17433], 00:10:11.668 | 99.00th=[30540], 99.50th=[39584], 99.90th=[43779], 99.95th=[47973], 00:10:11.668 | 99.99th=[47973] 00:10:11.668 write: IOPS=5576, BW=21.8MiB/s (22.8MB/s)(22.0MiB/1010msec); 0 zone resets 00:10:11.668 slat (nsec): min=1640, max=10981k, avg=74347.12, stdev=404759.22 00:10:11.668 clat (usec): min=2504, max=48056, avg=10745.79, stdev=4601.91 00:10:11.668 lat (usec): min=2512, max=48067, avg=10820.13, stdev=4632.03 00:10:11.668 clat percentiles (usec): 00:10:11.668 | 1.00th=[ 2966], 5.00th=[ 5407], 10.00th=[ 7308], 20.00th=[ 9241], 00:10:11.668 | 30.00th=[ 9634], 40.00th=[ 9765], 50.00th=[10028], 60.00th=[10159], 00:10:11.668 | 70.00th=[10945], 80.00th=[11207], 90.00th=[13304], 95.00th=[17433], 00:10:11.668 | 99.00th=[37487], 99.50th=[37487], 99.90th=[38011], 99.95th=[38011], 00:10:11.668 | 99.99th=[47973] 00:10:11.668 bw ( KiB/s): min=20944, max=24112, per=23.16%, avg=22528.00, stdev=2240.11, samples=2 00:10:11.668 iops : min= 5236, max= 6028, avg=5632.00, stdev=560.03, samples=2 00:10:11.668 lat (msec) : 4=1.42%, 10=37.14%, 20=58.61%, 50=2.83% 00:10:11.668 cpu : usr=4.06%, sys=5.05%, ctx=674, majf=0, minf=1 00:10:11.668 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:10:11.668 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:11.668 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:11.668 issued rwts: total=5613,5632,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:11.668 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:11.668 job2: (groupid=0, jobs=1): err= 0: pid=3709308: Tue Nov 26 19:06:24 2024 00:10:11.668 read: IOPS=3041, BW=11.9MiB/s (12.5MB/s)(12.0MiB/1010msec) 00:10:11.668 slat (nsec): min=925, max=31528k, avg=190737.98, stdev=1538043.77 00:10:11.668 clat (usec): min=7321, max=93290, avg=23174.67, stdev=20887.74 00:10:11.668 lat (usec): min=7390, max=93316, avg=23365.41, stdev=21053.47 00:10:11.668 clat percentiles (usec): 00:10:11.668 | 1.00th=[ 7635], 5.00th=[ 8291], 10.00th=[ 9372], 20.00th=[10421], 00:10:11.668 | 30.00th=[10814], 40.00th=[11731], 50.00th=[13435], 60.00th=[15008], 00:10:11.668 | 70.00th=[18482], 80.00th=[33424], 90.00th=[65274], 95.00th=[73925], 00:10:11.668 | 99.00th=[74974], 99.50th=[74974], 99.90th=[90702], 99.95th=[92799], 00:10:11.668 | 99.99th=[92799] 00:10:11.668 write: IOPS=3468, BW=13.5MiB/s (14.2MB/s)(13.7MiB/1010msec); 0 zone resets 00:10:11.668 slat (nsec): min=1715, max=15422k, avg=115674.99, stdev=762682.35 00:10:11.668 clat (usec): min=154, max=63779, avg=16139.03, stdev=11183.42 00:10:11.668 lat (usec): min=7219, max=63791, avg=16254.71, stdev=11228.56 00:10:11.668 clat percentiles (usec): 00:10:11.668 | 1.00th=[ 7963], 5.00th=[ 8717], 10.00th=[ 9110], 20.00th=[ 9503], 00:10:11.668 | 30.00th=[ 9765], 40.00th=[10159], 50.00th=[11469], 60.00th=[12387], 00:10:11.668 | 70.00th=[14222], 80.00th=[18744], 90.00th=[36439], 95.00th=[43254], 00:10:11.668 | 99.00th=[50594], 99.50th=[56886], 99.90th=[63701], 99.95th=[63701], 00:10:11.668 | 99.99th=[63701] 00:10:11.668 bw ( KiB/s): min=12784, max=14216, per=13.88%, avg=13500.00, stdev=1012.58, samples=2 00:10:11.668 iops : min= 3196, max= 3554, avg=3375.00, stdev=253.14, samples=2 00:10:11.668 lat (usec) : 250=0.02% 00:10:11.668 lat (msec) : 10=25.05%, 20=51.89%, 50=14.56%, 100=8.49% 00:10:11.668 cpu : usr=2.08%, sys=3.77%, ctx=354, majf=0, minf=1 00:10:11.668 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.5%, >=64=99.0% 00:10:11.668 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:11.668 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:11.668 issued rwts: total=3072,3503,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:11.668 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:11.668 job3: (groupid=0, jobs=1): err= 0: pid=3709321: Tue Nov 26 19:06:24 2024 00:10:11.668 read: IOPS=7376, BW=28.8MiB/s (30.2MB/s)(29.0MiB/1006msec) 00:10:11.668 slat (nsec): min=933, max=9022.2k, avg=68214.56, stdev=507772.23 00:10:11.668 clat (usec): min=2431, max=19138, avg=9092.30, stdev=2317.59 00:10:11.669 lat (usec): min=3616, max=19154, avg=9160.52, stdev=2346.10 00:10:11.669 clat percentiles (usec): 00:10:11.669 | 1.00th=[ 4490], 5.00th=[ 6194], 10.00th=[ 6652], 20.00th=[ 7373], 00:10:11.669 | 30.00th=[ 7701], 40.00th=[ 8094], 50.00th=[ 8455], 60.00th=[ 9241], 00:10:11.669 | 70.00th=[ 9896], 80.00th=[10683], 90.00th=[12387], 95.00th=[13829], 00:10:11.669 | 99.00th=[16450], 99.50th=[17171], 99.90th=[17957], 99.95th=[18482], 00:10:11.669 | 99.99th=[19268] 00:10:11.669 write: IOPS=7634, BW=29.8MiB/s (31.3MB/s)(30.0MiB/1006msec); 0 zone resets 00:10:11.669 slat (nsec): min=1649, max=8317.5k, avg=59025.16, stdev=425052.21 00:10:11.669 clat (usec): min=1358, max=18572, avg=7734.71, stdev=2154.89 00:10:11.669 lat (usec): min=1387, max=18574, avg=7793.73, stdev=2172.00 00:10:11.669 clat percentiles (usec): 00:10:11.669 | 1.00th=[ 3261], 5.00th=[ 4490], 10.00th=[ 4883], 20.00th=[ 5669], 00:10:11.669 | 30.00th=[ 6587], 40.00th=[ 7439], 50.00th=[ 7832], 60.00th=[ 8160], 00:10:11.669 | 70.00th=[ 8455], 80.00th=[ 9503], 90.00th=[10552], 95.00th=[10945], 00:10:11.669 | 99.00th=[13698], 99.50th=[13829], 99.90th=[16581], 99.95th=[17695], 00:10:11.669 | 99.99th=[18482] 00:10:11.669 bw ( KiB/s): min=29488, max=31952, per=31.58%, avg=30720.00, stdev=1742.31, samples=2 00:10:11.669 iops : min= 7372, max= 7988, avg=7680.00, stdev=435.58, samples=2 00:10:11.669 lat (msec) : 2=0.01%, 4=1.72%, 10=76.82%, 20=21.45% 00:10:11.669 cpu : usr=6.07%, sys=7.76%, ctx=521, majf=0, minf=1 00:10:11.669 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:10:11.669 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:11.669 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:11.669 issued rwts: total=7421,7680,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:11.669 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:11.669 00:10:11.669 Run status group 0 (all jobs): 00:10:11.669 READ: bw=92.0MiB/s (96.5MB/s), 11.9MiB/s-29.9MiB/s (12.5MB/s-31.4MB/s), io=92.9MiB (97.4MB), run=1002-1010msec 00:10:11.669 WRITE: bw=95.0MiB/s (99.6MB/s), 13.5MiB/s-30.2MiB/s (14.2MB/s-31.7MB/s), io=96.0MiB (101MB), run=1002-1010msec 00:10:11.669 00:10:11.669 Disk stats (read/write): 00:10:11.669 nvme0n1: ios=6194/6636, merge=0/0, ticks=41830/41474, in_queue=83304, util=92.18% 00:10:11.669 nvme0n2: ios=4135/4287, merge=0/0, ticks=48103/46314, in_queue=94417, util=85.27% 00:10:11.669 nvme0n3: ios=2101/2239, merge=0/0, ticks=17868/9568, in_queue=27436, util=91.53% 00:10:11.669 nvme0n4: ios=5685/6095, merge=0/0, ticks=47784/44054, in_queue=91838, util=93.01% 00:10:11.669 19:06:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:10:11.669 [global] 00:10:11.669 thread=1 00:10:11.669 invalidate=1 00:10:11.669 rw=randwrite 00:10:11.669 time_based=1 00:10:11.669 runtime=1 00:10:11.669 ioengine=libaio 00:10:11.669 direct=1 00:10:11.669 bs=4096 00:10:11.669 iodepth=128 00:10:11.669 norandommap=0 00:10:11.669 numjobs=1 00:10:11.669 00:10:11.669 verify_dump=1 00:10:11.669 verify_backlog=512 00:10:11.669 verify_state_save=0 00:10:11.669 do_verify=1 00:10:11.669 verify=crc32c-intel 00:10:11.669 [job0] 00:10:11.669 filename=/dev/nvme0n1 00:10:11.669 [job1] 00:10:11.669 filename=/dev/nvme0n2 00:10:11.669 [job2] 00:10:11.669 filename=/dev/nvme0n3 00:10:11.669 [job3] 00:10:11.669 filename=/dev/nvme0n4 00:10:11.669 Could not set queue depth (nvme0n1) 00:10:11.669 Could not set queue depth (nvme0n2) 00:10:11.669 Could not set queue depth (nvme0n3) 00:10:11.669 Could not set queue depth (nvme0n4) 00:10:11.930 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:11.930 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:11.930 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:11.930 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:10:11.930 fio-3.35 00:10:11.930 Starting 4 threads 00:10:13.317 00:10:13.317 job0: (groupid=0, jobs=1): err= 0: pid=3709817: Tue Nov 26 19:06:25 2024 00:10:13.317 read: IOPS=5581, BW=21.8MiB/s (22.9MB/s)(22.0MiB/1009msec) 00:10:13.317 slat (nsec): min=997, max=14922k, avg=93530.39, stdev=729063.63 00:10:13.317 clat (usec): min=3516, max=41829, avg=11433.42, stdev=3861.83 00:10:13.317 lat (usec): min=3522, max=41872, avg=11526.95, stdev=3930.47 00:10:13.317 clat percentiles (usec): 00:10:13.317 | 1.00th=[ 4555], 5.00th=[ 7832], 10.00th=[ 8848], 20.00th=[ 9634], 00:10:13.317 | 30.00th=[ 9765], 40.00th=[ 9896], 50.00th=[10159], 60.00th=[10421], 00:10:13.317 | 70.00th=[10814], 80.00th=[12911], 90.00th=[16450], 95.00th=[17957], 00:10:13.317 | 99.00th=[28705], 99.50th=[28705], 99.90th=[31851], 99.95th=[31851], 00:10:13.317 | 99.99th=[41681] 00:10:13.317 write: IOPS=5705, BW=22.3MiB/s (23.4MB/s)(22.5MiB/1009msec); 0 zone resets 00:10:13.317 slat (nsec): min=1671, max=13001k, avg=77239.51, stdev=429130.98 00:10:13.317 clat (usec): min=1170, max=61891, avg=10997.83, stdev=6721.94 00:10:13.317 lat (usec): min=1181, max=61899, avg=11075.06, stdev=6768.91 00:10:13.317 clat percentiles (usec): 00:10:13.317 | 1.00th=[ 3163], 5.00th=[ 5080], 10.00th=[ 6587], 20.00th=[ 8455], 00:10:13.317 | 30.00th=[ 9372], 40.00th=[ 9765], 50.00th=[ 9896], 60.00th=[10028], 00:10:13.317 | 70.00th=[10159], 80.00th=[11600], 90.00th=[12649], 95.00th=[19792], 00:10:13.317 | 99.00th=[47973], 99.50th=[51119], 99.90th=[60556], 99.95th=[62129], 00:10:13.317 | 99.99th=[62129] 00:10:13.317 bw ( KiB/s): min=19976, max=25136, per=24.43%, avg=22556.00, stdev=3648.67, samples=2 00:10:13.317 iops : min= 4994, max= 6284, avg=5639.00, stdev=912.17, samples=2 00:10:13.317 lat (msec) : 2=0.02%, 4=1.48%, 10=48.38%, 20=46.47%, 50=3.35% 00:10:13.317 lat (msec) : 100=0.32% 00:10:13.317 cpu : usr=4.17%, sys=4.86%, ctx=677, majf=0, minf=1 00:10:13.317 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:10:13.317 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:13.317 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:13.317 issued rwts: total=5632,5757,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:13.317 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:13.317 job1: (groupid=0, jobs=1): err= 0: pid=3709824: Tue Nov 26 19:06:25 2024 00:10:13.317 read: IOPS=3433, BW=13.4MiB/s (14.1MB/s)(13.5MiB/1005msec) 00:10:13.317 slat (nsec): min=919, max=51695k, avg=151986.21, stdev=1341186.30 00:10:13.317 clat (usec): min=4320, max=71186, avg=19212.49, stdev=13487.15 00:10:13.317 lat (usec): min=6268, max=78635, avg=19364.47, stdev=13579.04 00:10:13.317 clat percentiles (usec): 00:10:13.317 | 1.00th=[ 6915], 5.00th=[ 8848], 10.00th=[ 9372], 20.00th=[11469], 00:10:13.317 | 30.00th=[11863], 40.00th=[13698], 50.00th=[14746], 60.00th=[15795], 00:10:13.318 | 70.00th=[18744], 80.00th=[22414], 90.00th=[32637], 95.00th=[57410], 00:10:13.318 | 99.00th=[69731], 99.50th=[70779], 99.90th=[70779], 99.95th=[70779], 00:10:13.318 | 99.99th=[70779] 00:10:13.318 write: IOPS=3566, BW=13.9MiB/s (14.6MB/s)(14.0MiB/1005msec); 0 zone resets 00:10:13.318 slat (nsec): min=1564, max=18110k, avg=119538.94, stdev=829481.64 00:10:13.318 clat (usec): min=2578, max=67749, avg=16944.67, stdev=12464.17 00:10:13.318 lat (usec): min=2586, max=68159, avg=17064.21, stdev=12525.25 00:10:13.318 clat percentiles (usec): 00:10:13.318 | 1.00th=[ 4752], 5.00th=[ 6521], 10.00th=[ 6849], 20.00th=[ 7504], 00:10:13.318 | 30.00th=[ 8979], 40.00th=[11076], 50.00th=[13698], 60.00th=[15008], 00:10:13.318 | 70.00th=[17171], 80.00th=[20579], 90.00th=[36439], 95.00th=[46400], 00:10:13.318 | 99.00th=[61080], 99.50th=[67634], 99.90th=[67634], 99.95th=[67634], 00:10:13.318 | 99.99th=[67634] 00:10:13.318 bw ( KiB/s): min=12288, max=16384, per=15.53%, avg=14336.00, stdev=2896.31, samples=2 00:10:13.318 iops : min= 3072, max= 4096, avg=3584.00, stdev=724.08, samples=2 00:10:13.318 lat (msec) : 4=0.28%, 10=24.22%, 20=52.34%, 50=18.91%, 100=4.25% 00:10:13.318 cpu : usr=1.69%, sys=3.69%, ctx=411, majf=0, minf=1 00:10:13.318 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.5%, >=64=99.1% 00:10:13.318 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:13.318 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:13.318 issued rwts: total=3451,3584,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:13.318 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:13.318 job2: (groupid=0, jobs=1): err= 0: pid=3709833: Tue Nov 26 19:06:25 2024 00:10:13.318 read: IOPS=7104, BW=27.8MiB/s (29.1MB/s)(28.0MiB/1009msec) 00:10:13.318 slat (nsec): min=1012, max=8817.1k, avg=74191.27, stdev=561565.24 00:10:13.318 clat (usec): min=3451, max=23211, avg=9534.10, stdev=2422.88 00:10:13.318 lat (usec): min=3460, max=23430, avg=9608.29, stdev=2472.23 00:10:13.318 clat percentiles (usec): 00:10:13.318 | 1.00th=[ 5735], 5.00th=[ 6849], 10.00th=[ 7242], 20.00th=[ 7832], 00:10:13.318 | 30.00th=[ 8160], 40.00th=[ 8455], 50.00th=[ 8717], 60.00th=[ 9241], 00:10:13.318 | 70.00th=[ 9896], 80.00th=[11338], 90.00th=[13173], 95.00th=[14746], 00:10:13.318 | 99.00th=[16319], 99.50th=[16450], 99.90th=[18482], 99.95th=[20055], 00:10:13.318 | 99.99th=[23200] 00:10:13.318 write: IOPS=7228, BW=28.2MiB/s (29.6MB/s)(28.5MiB/1009msec); 0 zone resets 00:10:13.318 slat (nsec): min=1665, max=16640k, avg=58868.00, stdev=412929.16 00:10:13.318 clat (usec): min=712, max=18771, avg=8147.22, stdev=2196.47 00:10:13.318 lat (usec): min=724, max=18781, avg=8206.09, stdev=2217.72 00:10:13.318 clat percentiles (usec): 00:10:13.318 | 1.00th=[ 2900], 5.00th=[ 4948], 10.00th=[ 5800], 20.00th=[ 7177], 00:10:13.318 | 30.00th=[ 7635], 40.00th=[ 7832], 50.00th=[ 8094], 60.00th=[ 8356], 00:10:13.318 | 70.00th=[ 8586], 80.00th=[ 8979], 90.00th=[10028], 95.00th=[10683], 00:10:13.318 | 99.00th=[17957], 99.50th=[18744], 99.90th=[18744], 99.95th=[18744], 00:10:13.318 | 99.99th=[18744] 00:10:13.318 bw ( KiB/s): min=28672, max=28728, per=31.08%, avg=28700.00, stdev=39.60, samples=2 00:10:13.318 iops : min= 7168, max= 7182, avg=7175.00, stdev= 9.90, samples=2 00:10:13.318 lat (usec) : 750=0.02% 00:10:13.318 lat (msec) : 2=0.10%, 4=1.45%, 10=79.12%, 20=19.27%, 50=0.03% 00:10:13.318 cpu : usr=3.97%, sys=8.33%, ctx=655, majf=0, minf=1 00:10:13.318 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:10:13.318 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:13.318 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:13.318 issued rwts: total=7168,7294,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:13.318 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:13.318 job3: (groupid=0, jobs=1): err= 0: pid=3709840: Tue Nov 26 19:06:25 2024 00:10:13.318 read: IOPS=6578, BW=25.7MiB/s (26.9MB/s)(25.8MiB/1004msec) 00:10:13.318 slat (nsec): min=1039, max=8880.1k, avg=75402.25, stdev=549050.78 00:10:13.318 clat (usec): min=2762, max=25820, avg=9768.29, stdev=2723.57 00:10:13.318 lat (usec): min=2768, max=25824, avg=9843.69, stdev=2758.94 00:10:13.318 clat percentiles (usec): 00:10:13.318 | 1.00th=[ 5473], 5.00th=[ 6849], 10.00th=[ 7242], 20.00th=[ 7832], 00:10:13.318 | 30.00th=[ 8356], 40.00th=[ 8586], 50.00th=[ 9110], 60.00th=[ 9503], 00:10:13.318 | 70.00th=[10290], 80.00th=[11338], 90.00th=[13304], 95.00th=[14615], 00:10:13.318 | 99.00th=[19792], 99.50th=[22152], 99.90th=[23725], 99.95th=[25822], 00:10:13.318 | 99.99th=[25822] 00:10:13.318 write: IOPS=6629, BW=25.9MiB/s (27.2MB/s)(26.0MiB/1004msec); 0 zone resets 00:10:13.318 slat (nsec): min=1666, max=10516k, avg=67960.08, stdev=414097.15 00:10:13.318 clat (usec): min=826, max=25813, avg=9405.89, stdev=3790.75 00:10:13.318 lat (usec): min=839, max=25821, avg=9473.85, stdev=3814.12 00:10:13.318 clat percentiles (usec): 00:10:13.318 | 1.00th=[ 2474], 5.00th=[ 4555], 10.00th=[ 5014], 20.00th=[ 6063], 00:10:13.318 | 30.00th=[ 7308], 40.00th=[ 7832], 50.00th=[ 8160], 60.00th=[ 9503], 00:10:13.318 | 70.00th=[11338], 80.00th=[13960], 90.00th=[14877], 95.00th=[15270], 00:10:13.318 | 99.00th=[20579], 99.50th=[20579], 99.90th=[20841], 99.95th=[21890], 00:10:13.318 | 99.99th=[25822] 00:10:13.318 bw ( KiB/s): min=25448, max=27800, per=28.83%, avg=26624.00, stdev=1663.12, samples=2 00:10:13.318 iops : min= 6362, max= 6950, avg=6656.00, stdev=415.78, samples=2 00:10:13.318 lat (usec) : 1000=0.02% 00:10:13.318 lat (msec) : 2=0.35%, 4=1.52%, 10=63.05%, 20=34.05%, 50=1.00% 00:10:13.318 cpu : usr=4.89%, sys=6.98%, ctx=570, majf=0, minf=2 00:10:13.318 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.5% 00:10:13.318 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:13.318 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:10:13.318 issued rwts: total=6605,6656,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:13.318 latency : target=0, window=0, percentile=100.00%, depth=128 00:10:13.318 00:10:13.318 Run status group 0 (all jobs): 00:10:13.318 READ: bw=88.5MiB/s (92.8MB/s), 13.4MiB/s-27.8MiB/s (14.1MB/s-29.1MB/s), io=89.3MiB (93.6MB), run=1004-1009msec 00:10:13.318 WRITE: bw=90.2MiB/s (94.5MB/s), 13.9MiB/s-28.2MiB/s (14.6MB/s-29.6MB/s), io=91.0MiB (95.4MB), run=1004-1009msec 00:10:13.318 00:10:13.318 Disk stats (read/write): 00:10:13.318 nvme0n1: ios=5034/5120, merge=0/0, ticks=53396/46593, in_queue=99989, util=96.79% 00:10:13.318 nvme0n2: ios=2594/2838, merge=0/0, ticks=18757/15903, in_queue=34660, util=87.24% 00:10:13.318 nvme0n3: ios=6022/6144, merge=0/0, ticks=52810/48519, in_queue=101329, util=96.51% 00:10:13.318 nvme0n4: ios=5175/5479, merge=0/0, ticks=49552/52312, in_queue=101864, util=97.11% 00:10:13.318 19:06:25 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:10:13.318 19:06:25 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=3710133 00:10:13.318 19:06:25 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:10:13.318 19:06:25 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:10:13.318 [global] 00:10:13.318 thread=1 00:10:13.318 invalidate=1 00:10:13.318 rw=read 00:10:13.318 time_based=1 00:10:13.318 runtime=10 00:10:13.318 ioengine=libaio 00:10:13.318 direct=1 00:10:13.318 bs=4096 00:10:13.318 iodepth=1 00:10:13.318 norandommap=1 00:10:13.318 numjobs=1 00:10:13.318 00:10:13.318 [job0] 00:10:13.318 filename=/dev/nvme0n1 00:10:13.318 [job1] 00:10:13.318 filename=/dev/nvme0n2 00:10:13.318 [job2] 00:10:13.318 filename=/dev/nvme0n3 00:10:13.318 [job3] 00:10:13.318 filename=/dev/nvme0n4 00:10:13.318 Could not set queue depth (nvme0n1) 00:10:13.318 Could not set queue depth (nvme0n2) 00:10:13.318 Could not set queue depth (nvme0n3) 00:10:13.318 Could not set queue depth (nvme0n4) 00:10:13.577 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:13.577 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:13.577 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:13.577 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:10:13.577 fio-3.35 00:10:13.577 Starting 4 threads 00:10:16.880 19:06:28 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:10:16.880 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=1691648, buflen=4096 00:10:16.880 fio: pid=3710366, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:10:16.880 19:06:28 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:10:16.880 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=274432, buflen=4096 00:10:16.880 fio: pid=3710359, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:10:16.880 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:10:16.880 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:10:16.880 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:10:16.880 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:10:16.880 fio: io_u error on file /dev/nvme0n1: Operation not supported: read offset=6643712, buflen=4096 00:10:16.880 fio: pid=3710340, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:10:16.880 fio: io_u error on file /dev/nvme0n2: Operation not supported: read offset=2015232, buflen=4096 00:10:16.880 fio: pid=3710349, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:10:16.880 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:10:16.880 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:10:17.140 00:10:17.140 job0: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=3710340: Tue Nov 26 19:06:29 2024 00:10:17.140 read: IOPS=542, BW=2167KiB/s (2219kB/s)(6488KiB/2994msec) 00:10:17.140 slat (usec): min=6, max=11628, avg=36.16, stdev=301.43 00:10:17.140 clat (usec): min=613, max=42982, avg=1789.58, stdev=5744.89 00:10:17.140 lat (usec): min=640, max=53039, avg=1825.74, stdev=5817.04 00:10:17.140 clat percentiles (usec): 00:10:17.140 | 1.00th=[ 783], 5.00th=[ 865], 10.00th=[ 898], 20.00th=[ 938], 00:10:17.140 | 30.00th=[ 947], 40.00th=[ 955], 50.00th=[ 963], 60.00th=[ 971], 00:10:17.140 | 70.00th=[ 988], 80.00th=[ 1004], 90.00th=[ 1029], 95.00th=[ 1057], 00:10:17.140 | 99.00th=[41681], 99.50th=[42206], 99.90th=[42730], 99.95th=[42730], 00:10:17.140 | 99.99th=[42730] 00:10:17.140 bw ( KiB/s): min= 96, max= 4008, per=78.18%, avg=2577.60, stdev=1875.72, samples=5 00:10:17.140 iops : min= 24, max= 1002, avg=644.40, stdev=468.93, samples=5 00:10:17.140 lat (usec) : 750=0.49%, 1000=78.50% 00:10:17.140 lat (msec) : 2=18.92%, 50=2.03% 00:10:17.140 cpu : usr=1.40%, sys=1.77%, ctx=1625, majf=0, minf=1 00:10:17.140 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:17.140 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.140 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.140 issued rwts: total=1623,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:17.140 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:17.140 job1: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=3710349: Tue Nov 26 19:06:29 2024 00:10:17.140 read: IOPS=156, BW=625KiB/s (640kB/s)(1968KiB/3148msec) 00:10:17.140 slat (usec): min=6, max=13709, avg=94.41, stdev=823.33 00:10:17.140 clat (usec): min=281, max=42140, avg=6251.85, stdev=14047.35 00:10:17.140 lat (usec): min=307, max=42167, avg=6346.43, stdev=14045.63 00:10:17.140 clat percentiles (usec): 00:10:17.140 | 1.00th=[ 400], 5.00th=[ 498], 10.00th=[ 529], 20.00th=[ 644], 00:10:17.140 | 30.00th=[ 701], 40.00th=[ 758], 50.00th=[ 775], 60.00th=[ 791], 00:10:17.140 | 70.00th=[ 816], 80.00th=[ 857], 90.00th=[41681], 95.00th=[42206], 00:10:17.140 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:10:17.140 | 99.99th=[42206] 00:10:17.140 bw ( KiB/s): min= 88, max= 2378, per=15.87%, avg=523.00, stdev=916.14, samples=6 00:10:17.140 iops : min= 22, max= 594, avg=130.67, stdev=228.83, samples=6 00:10:17.140 lat (usec) : 500=5.07%, 750=31.03%, 1000=50.10% 00:10:17.140 lat (msec) : 10=0.20%, 50=13.39% 00:10:17.140 cpu : usr=0.32%, sys=0.25%, ctx=498, majf=0, minf=2 00:10:17.140 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:17.140 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.140 complete : 0=0.2%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.140 issued rwts: total=493,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:17.140 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:17.141 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=3710359: Tue Nov 26 19:06:29 2024 00:10:17.141 read: IOPS=24, BW=96.0KiB/s (98.3kB/s)(268KiB/2793msec) 00:10:17.141 slat (usec): min=25, max=212, avg=28.83, stdev=22.69 00:10:17.141 clat (usec): min=806, max=42769, avg=41286.61, stdev=5028.52 00:10:17.141 lat (usec): min=843, max=42982, avg=41315.47, stdev=5028.13 00:10:17.141 clat percentiles (usec): 00:10:17.141 | 1.00th=[ 807], 5.00th=[41157], 10.00th=[41681], 20.00th=[41681], 00:10:17.141 | 30.00th=[41681], 40.00th=[42206], 50.00th=[42206], 60.00th=[42206], 00:10:17.141 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:10:17.141 | 99.00th=[42730], 99.50th=[42730], 99.90th=[42730], 99.95th=[42730], 00:10:17.141 | 99.99th=[42730] 00:10:17.141 bw ( KiB/s): min= 96, max= 96, per=2.91%, avg=96.00, stdev= 0.00, samples=5 00:10:17.141 iops : min= 24, max= 24, avg=24.00, stdev= 0.00, samples=5 00:10:17.141 lat (usec) : 1000=1.47% 00:10:17.141 lat (msec) : 50=97.06% 00:10:17.141 cpu : usr=0.11%, sys=0.00%, ctx=69, majf=0, minf=2 00:10:17.141 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:17.141 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.141 complete : 0=1.4%, 4=98.6%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.141 issued rwts: total=68,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:17.141 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:17.141 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=3710366: Tue Nov 26 19:06:29 2024 00:10:17.141 read: IOPS=158, BW=634KiB/s (650kB/s)(1652KiB/2604msec) 00:10:17.141 slat (nsec): min=6537, max=35286, avg=25931.73, stdev=4918.34 00:10:17.141 clat (usec): min=578, max=43018, avg=6206.69, stdev=13731.29 00:10:17.141 lat (usec): min=605, max=43046, avg=6232.62, stdev=13731.79 00:10:17.141 clat percentiles (usec): 00:10:17.141 | 1.00th=[ 750], 5.00th=[ 824], 10.00th=[ 865], 20.00th=[ 906], 00:10:17.141 | 30.00th=[ 947], 40.00th=[ 955], 50.00th=[ 963], 60.00th=[ 979], 00:10:17.141 | 70.00th=[ 988], 80.00th=[ 1012], 90.00th=[41681], 95.00th=[42206], 00:10:17.141 | 99.00th=[42730], 99.50th=[43254], 99.90th=[43254], 99.95th=[43254], 00:10:17.141 | 99.99th=[43254] 00:10:17.141 bw ( KiB/s): min= 96, max= 1224, per=16.38%, avg=540.80, stdev=609.17, samples=5 00:10:17.141 iops : min= 24, max= 306, avg=135.20, stdev=152.29, samples=5 00:10:17.141 lat (usec) : 750=1.21%, 1000=72.95% 00:10:17.141 lat (msec) : 2=12.80%, 50=12.80% 00:10:17.141 cpu : usr=0.19%, sys=0.69%, ctx=414, majf=0, minf=2 00:10:17.141 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:10:17.141 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.141 complete : 0=0.2%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:10:17.141 issued rwts: total=414,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:10:17.141 latency : target=0, window=0, percentile=100.00%, depth=1 00:10:17.141 00:10:17.141 Run status group 0 (all jobs): 00:10:17.141 READ: bw=3296KiB/s (3375kB/s), 96.0KiB/s-2167KiB/s (98.3kB/s-2219kB/s), io=10.1MiB (10.6MB), run=2604-3148msec 00:10:17.141 00:10:17.141 Disk stats (read/write): 00:10:17.141 nvme0n1: ios=1618/0, merge=0/0, ticks=2669/0, in_queue=2669, util=94.26% 00:10:17.141 nvme0n2: ios=441/0, merge=0/0, ticks=3030/0, in_queue=3030, util=94.61% 00:10:17.141 nvme0n3: ios=62/0, merge=0/0, ticks=2558/0, in_queue=2558, util=96.03% 00:10:17.141 nvme0n4: ios=413/0, merge=0/0, ticks=2565/0, in_queue=2565, util=96.46% 00:10:17.141 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:10:17.141 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:10:17.401 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:10:17.401 19:06:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:10:17.663 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:10:17.663 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:10:17.663 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:10:17.663 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # wait 3710133 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:17.924 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1223 -- # local i=0 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1235 -- # return 0 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:10:17.924 nvmf hotplug test: fio failed as expected 00:10:17.924 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:18.185 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:10:18.186 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:18.186 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:18.186 rmmod nvme_tcp 00:10:18.186 rmmod nvme_fabrics 00:10:18.186 rmmod nvme_keyring 00:10:18.186 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@517 -- # '[' -n 3706614 ']' 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@518 -- # killprocess 3706614 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@954 -- # '[' -z 3706614 ']' 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@958 -- # kill -0 3706614 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@959 -- # uname 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3706614 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3706614' 00:10:18.447 killing process with pid 3706614 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@973 -- # kill 3706614 00:10:18.447 19:06:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@978 -- # wait 3706614 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-save 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-restore 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:18.447 19:06:31 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:21.025 00:10:21.025 real 0m30.211s 00:10:21.025 user 2m37.143s 00:10:21.025 sys 0m10.188s 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:10:21.025 ************************************ 00:10:21.025 END TEST nvmf_fio_target 00:10:21.025 ************************************ 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:21.025 ************************************ 00:10:21.025 START TEST nvmf_bdevio 00:10:21.025 ************************************ 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:10:21.025 * Looking for test storage... 00:10:21.025 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1693 -- # lcov --version 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:10:21.025 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:21.025 --rc genhtml_branch_coverage=1 00:10:21.025 --rc genhtml_function_coverage=1 00:10:21.025 --rc genhtml_legend=1 00:10:21.025 --rc geninfo_all_blocks=1 00:10:21.025 --rc geninfo_unexecuted_blocks=1 00:10:21.025 00:10:21.025 ' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:10:21.025 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:21.025 --rc genhtml_branch_coverage=1 00:10:21.025 --rc genhtml_function_coverage=1 00:10:21.025 --rc genhtml_legend=1 00:10:21.025 --rc geninfo_all_blocks=1 00:10:21.025 --rc geninfo_unexecuted_blocks=1 00:10:21.025 00:10:21.025 ' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:10:21.025 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:21.025 --rc genhtml_branch_coverage=1 00:10:21.025 --rc genhtml_function_coverage=1 00:10:21.025 --rc genhtml_legend=1 00:10:21.025 --rc geninfo_all_blocks=1 00:10:21.025 --rc geninfo_unexecuted_blocks=1 00:10:21.025 00:10:21.025 ' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:10:21.025 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:21.025 --rc genhtml_branch_coverage=1 00:10:21.025 --rc genhtml_function_coverage=1 00:10:21.025 --rc genhtml_legend=1 00:10:21.025 --rc geninfo_all_blocks=1 00:10:21.025 --rc geninfo_unexecuted_blocks=1 00:10:21.025 00:10:21.025 ' 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:10:21.025 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:21.026 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@476 -- # prepare_net_devs 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@438 -- # local -g is_hw=no 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # remove_spdk_ns 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:10:21.026 19:06:33 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:10:29.163 Found 0000:31:00.0 (0x8086 - 0x159b) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:10:29.163 Found 0000:31:00.1 (0x8086 - 0x159b) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:10:29.163 Found net devices under 0000:31:00.0: cvl_0_0 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:10:29.163 Found net devices under 0000:31:00.1: cvl_0_1 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # is_hw=yes 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:29.163 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:29.164 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:29.164 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.667 ms 00:10:29.164 00:10:29.164 --- 10.0.0.2 ping statistics --- 00:10:29.164 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:29.164 rtt min/avg/max/mdev = 0.667/0.667/0.667/0.000 ms 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:29.164 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:29.164 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.223 ms 00:10:29.164 00:10:29.164 --- 10.0.0.1 ping statistics --- 00:10:29.164 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:29.164 rtt min/avg/max/mdev = 0.223/0.223/0.223/0.000 ms 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@450 -- # return 0 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@726 -- # xtrace_disable 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@509 -- # nvmfpid=3716134 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@510 -- # waitforlisten 3716134 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x78 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@835 -- # '[' -z 3716134 ']' 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:29.164 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:29.164 19:06:41 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:29.425 [2024-11-26 19:06:41.845002] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:10:29.425 [2024-11-26 19:06:41.845069] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:29.425 [2024-11-26 19:06:41.958165] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:29.425 [2024-11-26 19:06:42.008829] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:29.425 [2024-11-26 19:06:42.008893] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:29.425 [2024-11-26 19:06:42.008902] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:29.425 [2024-11-26 19:06:42.008909] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:29.425 [2024-11-26 19:06:42.008916] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:29.425 [2024-11-26 19:06:42.010961] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:10:29.425 [2024-11-26 19:06:42.011289] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:10:29.425 [2024-11-26 19:06:42.011452] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:10:29.425 [2024-11-26 19:06:42.011454] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:10:30.367 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:30.367 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@868 -- # return 0 00:10:30.367 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:10:30.367 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@732 -- # xtrace_disable 00:10:30.367 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:30.368 [2024-11-26 19:06:42.712650] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:30.368 Malloc0 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:30.368 [2024-11-26 19:06:42.790898] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@560 -- # config=() 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@560 -- # local subsystem config 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:10:30.368 { 00:10:30.368 "params": { 00:10:30.368 "name": "Nvme$subsystem", 00:10:30.368 "trtype": "$TEST_TRANSPORT", 00:10:30.368 "traddr": "$NVMF_FIRST_TARGET_IP", 00:10:30.368 "adrfam": "ipv4", 00:10:30.368 "trsvcid": "$NVMF_PORT", 00:10:30.368 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:10:30.368 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:10:30.368 "hdgst": ${hdgst:-false}, 00:10:30.368 "ddgst": ${ddgst:-false} 00:10:30.368 }, 00:10:30.368 "method": "bdev_nvme_attach_controller" 00:10:30.368 } 00:10:30.368 EOF 00:10:30.368 )") 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@582 -- # cat 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@584 -- # jq . 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@585 -- # IFS=, 00:10:30.368 19:06:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:10:30.368 "params": { 00:10:30.368 "name": "Nvme1", 00:10:30.368 "trtype": "tcp", 00:10:30.368 "traddr": "10.0.0.2", 00:10:30.368 "adrfam": "ipv4", 00:10:30.368 "trsvcid": "4420", 00:10:30.368 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:10:30.368 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:10:30.368 "hdgst": false, 00:10:30.368 "ddgst": false 00:10:30.368 }, 00:10:30.368 "method": "bdev_nvme_attach_controller" 00:10:30.368 }' 00:10:30.368 [2024-11-26 19:06:42.848650] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:10:30.368 [2024-11-26 19:06:42.848720] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3716415 ] 00:10:30.368 [2024-11-26 19:06:42.935187] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:10:30.368 [2024-11-26 19:06:42.979164] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:30.368 [2024-11-26 19:06:42.979285] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:30.368 [2024-11-26 19:06:42.979288] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:30.628 I/O targets: 00:10:30.628 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:10:30.628 00:10:30.628 00:10:30.628 CUnit - A unit testing framework for C - Version 2.1-3 00:10:30.628 http://cunit.sourceforge.net/ 00:10:30.628 00:10:30.628 00:10:30.628 Suite: bdevio tests on: Nvme1n1 00:10:30.628 Test: blockdev write read block ...passed 00:10:30.628 Test: blockdev write zeroes read block ...passed 00:10:30.628 Test: blockdev write zeroes read no split ...passed 00:10:30.628 Test: blockdev write zeroes read split ...passed 00:10:30.628 Test: blockdev write zeroes read split partial ...passed 00:10:30.628 Test: blockdev reset ...[2024-11-26 19:06:43.241500] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:10:30.628 [2024-11-26 19:06:43.241560] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11d84b0 (9): Bad file descriptor 00:10:30.888 [2024-11-26 19:06:43.390579] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:10:30.888 passed 00:10:30.888 Test: blockdev write read 8 blocks ...passed 00:10:30.888 Test: blockdev write read size > 128k ...passed 00:10:30.888 Test: blockdev write read invalid size ...passed 00:10:30.888 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:10:30.888 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:10:30.888 Test: blockdev write read max offset ...passed 00:10:31.150 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:10:31.150 Test: blockdev writev readv 8 blocks ...passed 00:10:31.150 Test: blockdev writev readv 30 x 1block ...passed 00:10:31.150 Test: blockdev writev readv block ...passed 00:10:31.150 Test: blockdev writev readv size > 128k ...passed 00:10:31.150 Test: blockdev writev readv size > 128k in two iovs ...passed 00:10:31.150 Test: blockdev comparev and writev ...[2024-11-26 19:06:43.609506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.609533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.609545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.609551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.609937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.609947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.609957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.609963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.610322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.610332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.610342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.610348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.610720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.610729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.610740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:10:31.150 [2024-11-26 19:06:43.610750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:10:31.150 passed 00:10:31.150 Test: blockdev nvme passthru rw ...passed 00:10:31.150 Test: blockdev nvme passthru vendor specific ...[2024-11-26 19:06:43.693253] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:10:31.150 [2024-11-26 19:06:43.693266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.693487] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:10:31.150 [2024-11-26 19:06:43.693496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.693724] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:10:31.150 [2024-11-26 19:06:43.693733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:10:31.150 [2024-11-26 19:06:43.693944] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:10:31.150 [2024-11-26 19:06:43.693954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:10:31.150 passed 00:10:31.150 Test: blockdev nvme admin passthru ...passed 00:10:31.150 Test: blockdev copy ...passed 00:10:31.150 00:10:31.150 Run Summary: Type Total Ran Passed Failed Inactive 00:10:31.150 suites 1 1 n/a 0 0 00:10:31.150 tests 23 23 23 0 0 00:10:31.150 asserts 152 152 152 0 n/a 00:10:31.150 00:10:31.150 Elapsed time = 1.269 seconds 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@516 -- # nvmfcleanup 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:31.412 rmmod nvme_tcp 00:10:31.412 rmmod nvme_fabrics 00:10:31.412 rmmod nvme_keyring 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@517 -- # '[' -n 3716134 ']' 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@518 -- # killprocess 3716134 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@954 -- # '[' -z 3716134 ']' 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@958 -- # kill -0 3716134 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@959 -- # uname 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:31.412 19:06:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3716134 00:10:31.412 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@960 -- # process_name=reactor_3 00:10:31.412 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@964 -- # '[' reactor_3 = sudo ']' 00:10:31.412 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3716134' 00:10:31.412 killing process with pid 3716134 00:10:31.412 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@973 -- # kill 3716134 00:10:31.412 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@978 -- # wait 3716134 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-save 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-restore 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:31.674 19:06:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:34.240 00:10:34.240 real 0m13.103s 00:10:34.240 user 0m13.179s 00:10:34.240 sys 0m6.822s 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:10:34.240 ************************************ 00:10:34.240 END TEST nvmf_bdevio 00:10:34.240 ************************************ 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:10:34.240 00:10:34.240 real 5m15.718s 00:10:34.240 user 11m49.869s 00:10:34.240 sys 1m59.191s 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:34.240 ************************************ 00:10:34.240 END TEST nvmf_target_core 00:10:34.240 ************************************ 00:10:34.240 19:06:46 nvmf_tcp -- nvmf/nvmf.sh@15 -- # run_test nvmf_target_extra /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:10:34.240 19:06:46 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:34.240 19:06:46 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:34.240 19:06:46 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:10:34.240 ************************************ 00:10:34.240 START TEST nvmf_target_extra 00:10:34.240 ************************************ 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:10:34.240 * Looking for test storage... 00:10:34.240 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1693 -- # lcov --version 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # IFS=.-: 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # read -ra ver1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # IFS=.-: 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # read -ra ver2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@338 -- # local 'op=<' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@340 -- # ver1_l=2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@341 -- # ver2_l=1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@344 -- # case "$op" in 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@345 -- # : 1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # decimal 1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # ver1[v]=1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # decimal 2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # ver2[v]=2 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # return 0 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:10:34.240 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.240 --rc genhtml_branch_coverage=1 00:10:34.240 --rc genhtml_function_coverage=1 00:10:34.240 --rc genhtml_legend=1 00:10:34.240 --rc geninfo_all_blocks=1 00:10:34.240 --rc geninfo_unexecuted_blocks=1 00:10:34.240 00:10:34.240 ' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:10:34.240 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.240 --rc genhtml_branch_coverage=1 00:10:34.240 --rc genhtml_function_coverage=1 00:10:34.240 --rc genhtml_legend=1 00:10:34.240 --rc geninfo_all_blocks=1 00:10:34.240 --rc geninfo_unexecuted_blocks=1 00:10:34.240 00:10:34.240 ' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:10:34.240 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.240 --rc genhtml_branch_coverage=1 00:10:34.240 --rc genhtml_function_coverage=1 00:10:34.240 --rc genhtml_legend=1 00:10:34.240 --rc geninfo_all_blocks=1 00:10:34.240 --rc geninfo_unexecuted_blocks=1 00:10:34.240 00:10:34.240 ' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:10:34.240 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.240 --rc genhtml_branch_coverage=1 00:10:34.240 --rc genhtml_function_coverage=1 00:10:34.240 --rc genhtml_legend=1 00:10:34.240 --rc geninfo_all_blocks=1 00:10:34.240 --rc geninfo_unexecuted_blocks=1 00:10:34.240 00:10:34.240 ' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # uname -s 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@15 -- # shopt -s extglob 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:34.240 19:06:46 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- paths/export.sh@5 -- # export PATH 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@51 -- # : 0 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:34.241 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@13 -- # TEST_ARGS=("$@") 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@15 -- # [[ 0 -eq 0 ]] 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@16 -- # run_test nvmf_example /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:10:34.241 ************************************ 00:10:34.241 START TEST nvmf_example 00:10:34.241 ************************************ 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:10:34.241 * Looking for test storage... 00:10:34.241 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1693 -- # lcov --version 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # IFS=.-: 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # read -ra ver1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # IFS=.-: 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # read -ra ver2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@338 -- # local 'op=<' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@340 -- # ver1_l=2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@341 -- # ver2_l=1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@344 -- # case "$op" in 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@345 -- # : 1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # decimal 1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # ver1[v]=1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # decimal 2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # ver2[v]=2 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # return 0 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:10:34.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.241 --rc genhtml_branch_coverage=1 00:10:34.241 --rc genhtml_function_coverage=1 00:10:34.241 --rc genhtml_legend=1 00:10:34.241 --rc geninfo_all_blocks=1 00:10:34.241 --rc geninfo_unexecuted_blocks=1 00:10:34.241 00:10:34.241 ' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:10:34.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.241 --rc genhtml_branch_coverage=1 00:10:34.241 --rc genhtml_function_coverage=1 00:10:34.241 --rc genhtml_legend=1 00:10:34.241 --rc geninfo_all_blocks=1 00:10:34.241 --rc geninfo_unexecuted_blocks=1 00:10:34.241 00:10:34.241 ' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:10:34.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.241 --rc genhtml_branch_coverage=1 00:10:34.241 --rc genhtml_function_coverage=1 00:10:34.241 --rc genhtml_legend=1 00:10:34.241 --rc geninfo_all_blocks=1 00:10:34.241 --rc geninfo_unexecuted_blocks=1 00:10:34.241 00:10:34.241 ' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:10:34.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:34.241 --rc genhtml_branch_coverage=1 00:10:34.241 --rc genhtml_function_coverage=1 00:10:34.241 --rc genhtml_legend=1 00:10:34.241 --rc geninfo_all_blocks=1 00:10:34.241 --rc geninfo_unexecuted_blocks=1 00:10:34.241 00:10:34.241 ' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # uname -s 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@15 -- # shopt -s extglob 00:10:34.241 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@5 -- # export PATH 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@51 -- # : 0 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:34.242 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:34.242 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@11 -- # NVMF_EXAMPLE=("$SPDK_EXAMPLE_DIR/nvmf") 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@13 -- # MALLOC_BDEV_SIZE=64 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@24 -- # build_nvmf_example_args 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@17 -- # '[' 0 -eq 1 ']' 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@20 -- # NVMF_EXAMPLE+=(-i "$NVMF_APP_SHM_ID" -g 10000) 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@21 -- # NVMF_EXAMPLE+=("${NO_HUGE[@]}") 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@40 -- # timing_enter nvmf_example_test 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@726 -- # xtrace_disable 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@41 -- # nvmftestinit 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@476 -- # prepare_net_devs 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@438 -- # local -g is_hw=no 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # remove_spdk_ns 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@309 -- # xtrace_disable 00:10:34.503 19:06:46 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # pci_devs=() 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # net_devs=() 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # e810=() 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # local -ga e810 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # x722=() 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # local -ga x722 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # mlx=() 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # local -ga mlx 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:10:42.733 Found 0000:31:00.0 (0x8086 - 0x159b) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:10:42.733 Found 0000:31:00.1 (0x8086 - 0x159b) 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:42.733 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:10:42.734 Found net devices under 0000:31:00.0: cvl_0_0 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:10:42.734 Found net devices under 0000:31:00.1: cvl_0_1 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # is_hw=yes 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:42.734 19:06:54 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:42.734 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:42.734 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.619 ms 00:10:42.734 00:10:42.734 --- 10.0.0.2 ping statistics --- 00:10:42.734 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:42.734 rtt min/avg/max/mdev = 0.619/0.619/0.619/0.000 ms 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:42.734 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:42.734 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.304 ms 00:10:42.734 00:10:42.734 --- 10.0.0.1 ping statistics --- 00:10:42.734 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:42.734 rtt min/avg/max/mdev = 0.304/0.304/0.304/0.000 ms 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@450 -- # return 0 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@42 -- # nvmfexamplestart '-m 0xF' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@27 -- # timing_enter start_nvmf_example 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@726 -- # xtrace_disable 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@29 -- # '[' tcp == tcp ']' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@30 -- # NVMF_EXAMPLE=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_EXAMPLE[@]}") 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@34 -- # nvmfpid=3721500 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@35 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/nvmf -i 0 -g 10000 -m 0xF 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@36 -- # waitforlisten 3721500 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@835 -- # '[' -z 3721500 ']' 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:42.734 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:42.734 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@868 -- # return 0 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@37 -- # timing_exit start_nvmf_example 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@732 -- # xtrace_disable 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # rpc_cmd bdev_malloc_create 64 512 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # malloc_bdevs='Malloc0 ' 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@49 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@52 -- # for malloc_bdev in $malloc_bdevs 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@59 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:10:42.998 19:06:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:10:55.239 Initializing NVMe Controllers 00:10:55.239 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:55.239 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:10:55.239 Initialization complete. Launching workers. 00:10:55.239 ======================================================== 00:10:55.239 Latency(us) 00:10:55.239 Device Information : IOPS MiB/s Average min max 00:10:55.239 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 18205.49 71.12 3514.84 666.51 15752.92 00:10:55.239 ======================================================== 00:10:55.239 Total : 18205.49 71.12 3514.84 666.51 15752.92 00:10:55.239 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@65 -- # trap - SIGINT SIGTERM EXIT 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@66 -- # nvmftestfini 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@516 -- # nvmfcleanup 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@121 -- # sync 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@124 -- # set +e 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:55.239 rmmod nvme_tcp 00:10:55.239 rmmod nvme_fabrics 00:10:55.239 rmmod nvme_keyring 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@128 -- # set -e 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@129 -- # return 0 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@517 -- # '[' -n 3721500 ']' 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@518 -- # killprocess 3721500 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@954 -- # '[' -z 3721500 ']' 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@958 -- # kill -0 3721500 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@959 -- # uname 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3721500 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@960 -- # process_name=nvmf 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@964 -- # '[' nvmf = sudo ']' 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3721500' 00:10:55.239 killing process with pid 3721500 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@973 -- # kill 3721500 00:10:55.239 19:07:05 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@978 -- # wait 3721500 00:10:55.239 nvmf threads initialize successfully 00:10:55.239 bdev subsystem init successfully 00:10:55.239 created a nvmf target service 00:10:55.239 create targets's poll groups done 00:10:55.239 all subsystems of target started 00:10:55.239 nvmf target is running 00:10:55.239 all subsystems of target stopped 00:10:55.239 destroy targets's poll groups done 00:10:55.239 destroyed the nvmf target service 00:10:55.239 bdev subsystem finish successfully 00:10:55.239 nvmf threads destroy successfully 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@297 -- # iptr 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@791 -- # iptables-save 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@791 -- # iptables-restore 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:55.239 19:07:06 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@67 -- # timing_exit nvmf_example_test 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@732 -- # xtrace_disable 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:55.815 00:10:55.815 real 0m21.543s 00:10:55.815 user 0m44.297s 00:10:55.815 sys 0m7.425s 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:55.815 ************************************ 00:10:55.815 END TEST nvmf_example 00:10:55.815 ************************************ 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@17 -- # run_test nvmf_filesystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:10:55.815 ************************************ 00:10:55.815 START TEST nvmf_filesystem 00:10:55.815 ************************************ 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:10:55.815 * Looking for test storage... 00:10:55.815 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1693 -- # lcov --version 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:10:55.815 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:10:56.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.080 --rc genhtml_branch_coverage=1 00:10:56.080 --rc genhtml_function_coverage=1 00:10:56.080 --rc genhtml_legend=1 00:10:56.080 --rc geninfo_all_blocks=1 00:10:56.080 --rc geninfo_unexecuted_blocks=1 00:10:56.080 00:10:56.080 ' 00:10:56.080 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:10:56.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.080 --rc genhtml_branch_coverage=1 00:10:56.080 --rc genhtml_function_coverage=1 00:10:56.080 --rc genhtml_legend=1 00:10:56.080 --rc geninfo_all_blocks=1 00:10:56.081 --rc geninfo_unexecuted_blocks=1 00:10:56.081 00:10:56.081 ' 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:10:56.081 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.081 --rc genhtml_branch_coverage=1 00:10:56.081 --rc genhtml_function_coverage=1 00:10:56.081 --rc genhtml_legend=1 00:10:56.081 --rc geninfo_all_blocks=1 00:10:56.081 --rc geninfo_unexecuted_blocks=1 00:10:56.081 00:10:56.081 ' 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:10:56.081 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.081 --rc genhtml_branch_coverage=1 00:10:56.081 --rc genhtml_function_coverage=1 00:10:56.081 --rc genhtml_legend=1 00:10:56.081 --rc geninfo_all_blocks=1 00:10:56.081 --rc geninfo_unexecuted_blocks=1 00:10:56.081 00:10:56.081 ' 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@34 -- # set -e 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@36 -- # shopt -s extglob 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@37 -- # shopt -s inherit_errexit 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@39 -- # '[' -z /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output ']' 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@44 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh ]] 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@17 -- # CONFIG_MAX_NUMA_NODES=1 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@18 -- # CONFIG_PGO_CAPTURE=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@19 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@20 -- # CONFIG_ENV=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@21 -- # CONFIG_LTO=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@22 -- # CONFIG_ISCSI_INITIATOR=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@23 -- # CONFIG_CET=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@24 -- # CONFIG_VBDEV_COMPRESS_MLX5=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@25 -- # CONFIG_OCF_PATH= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@26 -- # CONFIG_RDMA_SET_TOS=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@27 -- # CONFIG_AIO_FSDEV=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@28 -- # CONFIG_HAVE_ARC4RANDOM=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@29 -- # CONFIG_HAVE_LIBARCHIVE=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@30 -- # CONFIG_UBLK=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@31 -- # CONFIG_ISAL_CRYPTO=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@32 -- # CONFIG_OPENSSL_PATH= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@33 -- # CONFIG_OCF=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@34 -- # CONFIG_FUSE=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@35 -- # CONFIG_VTUNE_DIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@36 -- # CONFIG_FUZZER_LIB= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@37 -- # CONFIG_FUZZER=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@38 -- # CONFIG_FSDEV=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@39 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@40 -- # CONFIG_CRYPTO=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@41 -- # CONFIG_PGO_USE=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@42 -- # CONFIG_VHOST=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@43 -- # CONFIG_DAOS=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@44 -- # CONFIG_DPDK_INC_DIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@45 -- # CONFIG_DAOS_DIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@46 -- # CONFIG_UNIT_TESTS=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@47 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@48 -- # CONFIG_VIRTIO=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@49 -- # CONFIG_DPDK_UADK=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@50 -- # CONFIG_COVERAGE=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@51 -- # CONFIG_RDMA=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@52 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIM=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@53 -- # CONFIG_HAVE_LZ4=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@54 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@55 -- # CONFIG_URING_PATH= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@56 -- # CONFIG_XNVME=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@57 -- # CONFIG_VFIO_USER=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@58 -- # CONFIG_ARCH=native 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@59 -- # CONFIG_HAVE_EVP_MAC=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@60 -- # CONFIG_URING_ZNS=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@61 -- # CONFIG_WERROR=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@62 -- # CONFIG_HAVE_LIBBSD=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@63 -- # CONFIG_UBSAN=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@64 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@65 -- # CONFIG_IPSEC_MB_DIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@66 -- # CONFIG_GOLANG=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@67 -- # CONFIG_ISAL=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@68 -- # CONFIG_IDXD_KERNEL=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@69 -- # CONFIG_DPDK_LIB_DIR= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@70 -- # CONFIG_RDMA_PROV=verbs 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@71 -- # CONFIG_APPS=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@72 -- # CONFIG_SHARED=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@73 -- # CONFIG_HAVE_KEYUTILS=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@74 -- # CONFIG_FC_PATH= 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@75 -- # CONFIG_DPDK_PKG_CONFIG=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@76 -- # CONFIG_FC=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@77 -- # CONFIG_AVAHI=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@78 -- # CONFIG_FIO_PLUGIN=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@79 -- # CONFIG_RAID5F=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@80 -- # CONFIG_EXAMPLES=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@81 -- # CONFIG_TESTS=y 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@82 -- # CONFIG_CRYPTO_MLX5=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@83 -- # CONFIG_MAX_LCORES=128 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@84 -- # CONFIG_IPSEC_MB=n 00:10:56.081 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@85 -- # CONFIG_PGO_DIR= 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@86 -- # CONFIG_DEBUG=y 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@87 -- # CONFIG_DPDK_COMPRESSDEV=n 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@88 -- # CONFIG_CROSS_PREFIX= 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@89 -- # CONFIG_COPY_FILE_RANGE=y 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@90 -- # CONFIG_URING=n 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/config.h ]] 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:10:56.082 #define SPDK_CONFIG_H 00:10:56.082 #define SPDK_CONFIG_AIO_FSDEV 1 00:10:56.082 #define SPDK_CONFIG_APPS 1 00:10:56.082 #define SPDK_CONFIG_ARCH native 00:10:56.082 #undef SPDK_CONFIG_ASAN 00:10:56.082 #undef SPDK_CONFIG_AVAHI 00:10:56.082 #undef SPDK_CONFIG_CET 00:10:56.082 #define SPDK_CONFIG_COPY_FILE_RANGE 1 00:10:56.082 #define SPDK_CONFIG_COVERAGE 1 00:10:56.082 #define SPDK_CONFIG_CROSS_PREFIX 00:10:56.082 #undef SPDK_CONFIG_CRYPTO 00:10:56.082 #undef SPDK_CONFIG_CRYPTO_MLX5 00:10:56.082 #undef SPDK_CONFIG_CUSTOMOCF 00:10:56.082 #undef SPDK_CONFIG_DAOS 00:10:56.082 #define SPDK_CONFIG_DAOS_DIR 00:10:56.082 #define SPDK_CONFIG_DEBUG 1 00:10:56.082 #undef SPDK_CONFIG_DPDK_COMPRESSDEV 00:10:56.082 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:10:56.082 #define SPDK_CONFIG_DPDK_INC_DIR 00:10:56.082 #define SPDK_CONFIG_DPDK_LIB_DIR 00:10:56.082 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:10:56.082 #undef SPDK_CONFIG_DPDK_UADK 00:10:56.082 #define SPDK_CONFIG_ENV /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:10:56.082 #define SPDK_CONFIG_EXAMPLES 1 00:10:56.082 #undef SPDK_CONFIG_FC 00:10:56.082 #define SPDK_CONFIG_FC_PATH 00:10:56.082 #define SPDK_CONFIG_FIO_PLUGIN 1 00:10:56.082 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:10:56.082 #define SPDK_CONFIG_FSDEV 1 00:10:56.082 #undef SPDK_CONFIG_FUSE 00:10:56.082 #undef SPDK_CONFIG_FUZZER 00:10:56.082 #define SPDK_CONFIG_FUZZER_LIB 00:10:56.082 #undef SPDK_CONFIG_GOLANG 00:10:56.082 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:10:56.082 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:10:56.082 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:10:56.082 #define SPDK_CONFIG_HAVE_KEYUTILS 1 00:10:56.082 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:10:56.082 #undef SPDK_CONFIG_HAVE_LIBBSD 00:10:56.082 #undef SPDK_CONFIG_HAVE_LZ4 00:10:56.082 #define SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIM 1 00:10:56.082 #undef SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC 00:10:56.082 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:10:56.082 #define SPDK_CONFIG_IDXD 1 00:10:56.082 #define SPDK_CONFIG_IDXD_KERNEL 1 00:10:56.082 #undef SPDK_CONFIG_IPSEC_MB 00:10:56.082 #define SPDK_CONFIG_IPSEC_MB_DIR 00:10:56.082 #define SPDK_CONFIG_ISAL 1 00:10:56.082 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:10:56.082 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:10:56.082 #define SPDK_CONFIG_LIBDIR 00:10:56.082 #undef SPDK_CONFIG_LTO 00:10:56.082 #define SPDK_CONFIG_MAX_LCORES 128 00:10:56.082 #define SPDK_CONFIG_MAX_NUMA_NODES 1 00:10:56.082 #define SPDK_CONFIG_NVME_CUSE 1 00:10:56.082 #undef SPDK_CONFIG_OCF 00:10:56.082 #define SPDK_CONFIG_OCF_PATH 00:10:56.082 #define SPDK_CONFIG_OPENSSL_PATH 00:10:56.082 #undef SPDK_CONFIG_PGO_CAPTURE 00:10:56.082 #define SPDK_CONFIG_PGO_DIR 00:10:56.082 #undef SPDK_CONFIG_PGO_USE 00:10:56.082 #define SPDK_CONFIG_PREFIX /usr/local 00:10:56.082 #undef SPDK_CONFIG_RAID5F 00:10:56.082 #undef SPDK_CONFIG_RBD 00:10:56.082 #define SPDK_CONFIG_RDMA 1 00:10:56.082 #define SPDK_CONFIG_RDMA_PROV verbs 00:10:56.082 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:10:56.082 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:10:56.082 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:10:56.082 #define SPDK_CONFIG_SHARED 1 00:10:56.082 #undef SPDK_CONFIG_SMA 00:10:56.082 #define SPDK_CONFIG_TESTS 1 00:10:56.082 #undef SPDK_CONFIG_TSAN 00:10:56.082 #define SPDK_CONFIG_UBLK 1 00:10:56.082 #define SPDK_CONFIG_UBSAN 1 00:10:56.082 #undef SPDK_CONFIG_UNIT_TESTS 00:10:56.082 #undef SPDK_CONFIG_URING 00:10:56.082 #define SPDK_CONFIG_URING_PATH 00:10:56.082 #undef SPDK_CONFIG_URING_ZNS 00:10:56.082 #undef SPDK_CONFIG_USDT 00:10:56.082 #undef SPDK_CONFIG_VBDEV_COMPRESS 00:10:56.082 #undef SPDK_CONFIG_VBDEV_COMPRESS_MLX5 00:10:56.082 #define SPDK_CONFIG_VFIO_USER 1 00:10:56.082 #define SPDK_CONFIG_VFIO_USER_DIR 00:10:56.082 #define SPDK_CONFIG_VHOST 1 00:10:56.082 #define SPDK_CONFIG_VIRTIO 1 00:10:56.082 #undef SPDK_CONFIG_VTUNE 00:10:56.082 #define SPDK_CONFIG_VTUNE_DIR 00:10:56.082 #define SPDK_CONFIG_WERROR 1 00:10:56.082 #define SPDK_CONFIG_WPDK_DIR 00:10:56.082 #undef SPDK_CONFIG_XNVME 00:10:56.082 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@56 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/../../../ 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@64 -- # TEST_TAG=N/A 00:10:56.082 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@65 -- # TEST_TAG_FILE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.run_test_name 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@67 -- # PM_OUTPUTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # uname -s 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # PM_OS=Linux 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # MONITOR_RESOURCES_SUDO=() 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # declare -A MONITOR_RESOURCES_SUDO 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@71 -- # MONITOR_RESOURCES_SUDO["collect-bmc-pm"]=1 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@72 -- # MONITOR_RESOURCES_SUDO["collect-cpu-load"]=0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@73 -- # MONITOR_RESOURCES_SUDO["collect-cpu-temp"]=0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@74 -- # MONITOR_RESOURCES_SUDO["collect-vmstat"]=0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[0]= 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[1]='sudo -E' 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@78 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@79 -- # [[ Linux == FreeBSD ]] 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ Linux == Linux ]] 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ............................... != QEMU ]] 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ! -e /.dockerenv ]] 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@84 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@85 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@88 -- # [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power ]] 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@58 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@59 -- # export RUN_NIGHTLY 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@62 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@63 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@64 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@65 -- # export SPDK_RUN_VALGRIND 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@66 -- # : 1 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@67 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@68 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@69 -- # export SPDK_TEST_UNITTEST 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@70 -- # : 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@71 -- # export SPDK_TEST_AUTOBUILD 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@72 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@73 -- # export SPDK_TEST_RELEASE_BUILD 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@74 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@75 -- # export SPDK_TEST_ISAL 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@76 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@77 -- # export SPDK_TEST_ISCSI 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@78 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@79 -- # export SPDK_TEST_ISCSI_INITIATOR 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@80 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@81 -- # export SPDK_TEST_NVME 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@82 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@83 -- # export SPDK_TEST_NVME_PMR 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@84 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@85 -- # export SPDK_TEST_NVME_BP 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@86 -- # : 1 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@87 -- # export SPDK_TEST_NVME_CLI 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@88 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@89 -- # export SPDK_TEST_NVME_CUSE 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@90 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@91 -- # export SPDK_TEST_NVME_FDP 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@92 -- # : 1 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@93 -- # export SPDK_TEST_NVMF 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@94 -- # : 1 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@95 -- # export SPDK_TEST_VFIOUSER 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@96 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@97 -- # export SPDK_TEST_VFIOUSER_QEMU 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@98 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@99 -- # export SPDK_TEST_FUZZER 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@100 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@101 -- # export SPDK_TEST_FUZZER_SHORT 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@102 -- # : tcp 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@103 -- # export SPDK_TEST_NVMF_TRANSPORT 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@104 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@105 -- # export SPDK_TEST_RBD 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@106 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@107 -- # export SPDK_TEST_VHOST 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@108 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@109 -- # export SPDK_TEST_BLOCKDEV 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@110 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@111 -- # export SPDK_TEST_RAID 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@112 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@113 -- # export SPDK_TEST_IOAT 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@114 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@115 -- # export SPDK_TEST_BLOBFS 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@116 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@117 -- # export SPDK_TEST_VHOST_INIT 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@118 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@119 -- # export SPDK_TEST_LVOL 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@120 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@121 -- # export SPDK_TEST_VBDEV_COMPRESS 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@122 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@123 -- # export SPDK_RUN_ASAN 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@124 -- # : 1 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@125 -- # export SPDK_RUN_UBSAN 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@126 -- # : 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@127 -- # export SPDK_RUN_EXTERNAL_DPDK 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@128 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@129 -- # export SPDK_RUN_NON_ROOT 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@130 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@131 -- # export SPDK_TEST_CRYPTO 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@132 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@133 -- # export SPDK_TEST_FTL 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@134 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@135 -- # export SPDK_TEST_OCF 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@136 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@137 -- # export SPDK_TEST_VMD 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@138 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@139 -- # export SPDK_TEST_OPAL 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@140 -- # : 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@141 -- # export SPDK_TEST_NATIVE_DPDK 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@142 -- # : true 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@143 -- # export SPDK_AUTOTEST_X 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@144 -- # : 0 00:10:56.083 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@145 -- # export SPDK_TEST_URING 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@146 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@147 -- # export SPDK_TEST_USDT 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@148 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@149 -- # export SPDK_TEST_USE_IGB_UIO 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@150 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@151 -- # export SPDK_TEST_SCHEDULER 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@152 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@153 -- # export SPDK_TEST_SCANBUILD 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@154 -- # : e810 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@155 -- # export SPDK_TEST_NVMF_NICS 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@156 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@157 -- # export SPDK_TEST_SMA 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@158 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@159 -- # export SPDK_TEST_DAOS 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@160 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@161 -- # export SPDK_TEST_XNVME 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@162 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@163 -- # export SPDK_TEST_ACCEL 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@164 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@165 -- # export SPDK_TEST_ACCEL_DSA 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@166 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@167 -- # export SPDK_TEST_ACCEL_IAA 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@169 -- # : 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@170 -- # export SPDK_TEST_FUZZER_TARGET 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@171 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@172 -- # export SPDK_TEST_NVMF_MDNS 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@173 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@174 -- # export SPDK_JSONRPC_GO_CLIENT 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@175 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@176 -- # export SPDK_TEST_SETUP 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@177 -- # : 0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@178 -- # export SPDK_TEST_NVME_INTERRUPT 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@183 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@183 -- # VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@184 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@184 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@187 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@187 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@191 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@191 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@195 -- # export PYTHONDONTWRITEBYTECODE=1 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@195 -- # PYTHONDONTWRITEBYTECODE=1 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@199 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@199 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@200 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@200 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@204 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@205 -- # rm -rf /var/tmp/asan_suppression_file 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@206 -- # cat 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@242 -- # echo leak:libfuse3.so 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@246 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@246 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@248 -- # '[' -z /var/spdk/dependencies ']' 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@251 -- # export DEPENDENCY_DIR 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@255 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@255 -- # SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@256 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@256 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@259 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@259 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@262 -- # export AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@262 -- # AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@265 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@265 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@267 -- # _LCOV_MAIN=0 00:10:56.084 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@268 -- # _LCOV_LLVM=1 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@269 -- # _LCOV= 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@270 -- # [[ '' == *clang* ]] 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@270 -- # [[ 0 -eq 1 ]] 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@272 -- # _lcov_opt[_LCOV_LLVM]='--gcov-tool /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/fuzz/llvm/llvm-gcov.sh' 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@273 -- # _lcov_opt[_LCOV_MAIN]= 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@275 -- # lcov_opt= 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@278 -- # '[' 0 -eq 0 ']' 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@279 -- # export valgrind= 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@279 -- # valgrind= 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # uname -s 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # '[' Linux = Linux ']' 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@286 -- # HUGEMEM=4096 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@287 -- # export CLEAR_HUGE=yes 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@287 -- # CLEAR_HUGE=yes 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@289 -- # MAKE=make 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@290 -- # MAKEFLAGS=-j144 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@306 -- # export HUGEMEM=4096 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@306 -- # HUGEMEM=4096 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@308 -- # NO_HUGE=() 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@309 -- # TEST_MODE= 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@310 -- # for i in "$@" 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@311 -- # case "$i" in 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@316 -- # TEST_TRANSPORT=tcp 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@331 -- # [[ -z 3724282 ]] 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@331 -- # kill -0 3724282 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1678 -- # set_test_storage 2147483648 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@341 -- # [[ -v testdir ]] 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@343 -- # local requested_size=2147483648 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@344 -- # local mount target_dir 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@346 -- # local -A mounts fss sizes avails uses 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@347 -- # local source fs size avail mount use 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@349 -- # local storage_fallback storage_candidates 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@351 -- # mktemp -udt spdk.XXXXXX 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@351 -- # storage_fallback=/tmp/spdk.Oz62sn 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@356 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@358 -- # [[ -n '' ]] 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@363 -- # [[ -n '' ]] 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@368 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target /tmp/spdk.Oz62sn/tests/target /tmp/spdk.Oz62sn 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # requested_size=2214592512 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@340 -- # df -T 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@340 -- # grep -v Filesystem 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=spdk_devtmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=devtmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=67108864 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=67108864 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=0 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=/dev/pmem0 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=ext2 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=4096 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=5284429824 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=5284425728 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=spdk_root 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=overlay 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=122260680704 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=129356550144 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=7095869440 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=64666906624 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=64678273024 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=11366400 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=25847697408 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=25871310848 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=23613440 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=efivarfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=efivarfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=175104 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=507904 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=328704 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=64677548032 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=64678277120 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=729088 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.085 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=12935639040 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=12935651328 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=12288 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@379 -- # printf '* Looking for test storage...\n' 00:10:56.086 * Looking for test storage... 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@381 -- # local target_space new_size 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@382 -- # for target_dir in "${storage_candidates[@]}" 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # df /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # awk '$1 !~ /Filesystem/{print $6}' 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # mount=/ 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@387 -- # target_space=122260680704 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@388 -- # (( target_space == 0 || target_space < requested_size )) 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # (( target_space >= requested_size )) 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # [[ overlay == tmpfs ]] 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # [[ overlay == ramfs ]] 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # [[ / == / ]] 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@394 -- # new_size=9310461952 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@395 -- # (( new_size * 100 / sizes[/] > 95 )) 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@400 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@400 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@401 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:56.086 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@402 -- # return 0 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1680 -- # set -o errtrace 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1681 -- # shopt -s extdebug 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1682 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1684 -- # PS4=' \t ${test_domain:-} -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1685 -- # true 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1687 -- # xtrace_fd 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -n 15 ]] 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/15 ]] 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@27 -- # exec 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@29 -- # exec 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@31 -- # xtrace_restore 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@18 -- # set -x 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1693 -- # lcov --version 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:56.086 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:10:56.349 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.349 --rc genhtml_branch_coverage=1 00:10:56.349 --rc genhtml_function_coverage=1 00:10:56.349 --rc genhtml_legend=1 00:10:56.349 --rc geninfo_all_blocks=1 00:10:56.349 --rc geninfo_unexecuted_blocks=1 00:10:56.349 00:10:56.349 ' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:10:56.349 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.349 --rc genhtml_branch_coverage=1 00:10:56.349 --rc genhtml_function_coverage=1 00:10:56.349 --rc genhtml_legend=1 00:10:56.349 --rc geninfo_all_blocks=1 00:10:56.349 --rc geninfo_unexecuted_blocks=1 00:10:56.349 00:10:56.349 ' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:10:56.349 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.349 --rc genhtml_branch_coverage=1 00:10:56.349 --rc genhtml_function_coverage=1 00:10:56.349 --rc genhtml_legend=1 00:10:56.349 --rc geninfo_all_blocks=1 00:10:56.349 --rc geninfo_unexecuted_blocks=1 00:10:56.349 00:10:56.349 ' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:10:56.349 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.349 --rc genhtml_branch_coverage=1 00:10:56.349 --rc genhtml_function_coverage=1 00:10:56.349 --rc genhtml_legend=1 00:10:56.349 --rc geninfo_all_blocks=1 00:10:56.349 --rc geninfo_unexecuted_blocks=1 00:10:56.349 00:10:56.349 ' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # uname -s 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@51 -- # : 0 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:56.349 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@12 -- # MALLOC_BDEV_SIZE=512 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@15 -- # nvmftestinit 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@476 -- # prepare_net_devs 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@438 -- # local -g is_hw=no 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # remove_spdk_ns 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@309 -- # xtrace_disable 00:10:56.349 19:07:08 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:11:04.496 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:04.496 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # pci_devs=() 00:11:04.496 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:04.496 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:04.496 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:04.496 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # net_devs=() 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # e810=() 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # local -ga e810 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # x722=() 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # local -ga x722 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # mlx=() 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # local -ga mlx 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:11:04.497 Found 0000:31:00.0 (0x8086 - 0x159b) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:11:04.497 Found 0000:31:00.1 (0x8086 - 0x159b) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:11:04.497 Found net devices under 0000:31:00.0: cvl_0_0 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:11:04.497 Found net devices under 0000:31:00.1: cvl_0_1 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # is_hw=yes 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:04.497 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:04.498 19:07:16 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:04.498 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:04.498 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.706 ms 00:11:04.498 00:11:04.498 --- 10.0.0.2 ping statistics --- 00:11:04.498 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:04.498 rtt min/avg/max/mdev = 0.706/0.706/0.706/0.000 ms 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:04.498 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:04.498 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.322 ms 00:11:04.498 00:11:04.498 --- 10.0.0.1 ping statistics --- 00:11:04.498 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:04.498 rtt min/avg/max/mdev = 0.322/0.322/0.322/0.000 ms 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@450 -- # return 0 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@105 -- # run_test nvmf_filesystem_no_in_capsule nvmf_filesystem_part 0 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:04.498 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:11:04.758 ************************************ 00:11:04.758 START TEST nvmf_filesystem_no_in_capsule 00:11:04.758 ************************************ 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1129 -- # nvmf_filesystem_part 0 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@47 -- # in_capsule=0 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@509 -- # nvmfpid=3728597 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@510 -- # waitforlisten 3728597 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@835 -- # '[' -z 3728597 ']' 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:04.758 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:04.758 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:04.758 [2024-11-26 19:07:17.191701] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:11:04.758 [2024-11-26 19:07:17.191749] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:04.758 [2024-11-26 19:07:17.282397] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:04.758 [2024-11-26 19:07:17.322406] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:04.758 [2024-11-26 19:07:17.322446] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:04.758 [2024-11-26 19:07:17.322454] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:04.758 [2024-11-26 19:07:17.322461] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:04.758 [2024-11-26 19:07:17.322467] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:04.758 [2024-11-26 19:07:17.324015] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:04.758 [2024-11-26 19:07:17.324132] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:04.758 [2024-11-26 19:07:17.324282] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:04.758 [2024-11-26 19:07:17.324283] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:05.699 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:05.699 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@868 -- # return 0 00:11:05.699 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:11:05.699 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:05.699 19:07:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:05.699 [2024-11-26 19:07:18.035515] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:05.699 Malloc1 00:11:05.699 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:05.700 [2024-11-26 19:07:18.167077] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # local bdev_name=Malloc1 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # local bdev_info 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # local bs 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1385 -- # local nb 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1386 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1386 -- # bdev_info='[ 00:11:05.700 { 00:11:05.700 "name": "Malloc1", 00:11:05.700 "aliases": [ 00:11:05.700 "001ceedd-e3ed-412a-9cee-605b3880f364" 00:11:05.700 ], 00:11:05.700 "product_name": "Malloc disk", 00:11:05.700 "block_size": 512, 00:11:05.700 "num_blocks": 1048576, 00:11:05.700 "uuid": "001ceedd-e3ed-412a-9cee-605b3880f364", 00:11:05.700 "assigned_rate_limits": { 00:11:05.700 "rw_ios_per_sec": 0, 00:11:05.700 "rw_mbytes_per_sec": 0, 00:11:05.700 "r_mbytes_per_sec": 0, 00:11:05.700 "w_mbytes_per_sec": 0 00:11:05.700 }, 00:11:05.700 "claimed": true, 00:11:05.700 "claim_type": "exclusive_write", 00:11:05.700 "zoned": false, 00:11:05.700 "supported_io_types": { 00:11:05.700 "read": true, 00:11:05.700 "write": true, 00:11:05.700 "unmap": true, 00:11:05.700 "flush": true, 00:11:05.700 "reset": true, 00:11:05.700 "nvme_admin": false, 00:11:05.700 "nvme_io": false, 00:11:05.700 "nvme_io_md": false, 00:11:05.700 "write_zeroes": true, 00:11:05.700 "zcopy": true, 00:11:05.700 "get_zone_info": false, 00:11:05.700 "zone_management": false, 00:11:05.700 "zone_append": false, 00:11:05.700 "compare": false, 00:11:05.700 "compare_and_write": false, 00:11:05.700 "abort": true, 00:11:05.700 "seek_hole": false, 00:11:05.700 "seek_data": false, 00:11:05.700 "copy": true, 00:11:05.700 "nvme_iov_md": false 00:11:05.700 }, 00:11:05.700 "memory_domains": [ 00:11:05.700 { 00:11:05.700 "dma_device_id": "system", 00:11:05.700 "dma_device_type": 1 00:11:05.700 }, 00:11:05.700 { 00:11:05.700 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:05.700 "dma_device_type": 2 00:11:05.700 } 00:11:05.700 ], 00:11:05.700 "driver_specific": {} 00:11:05.700 } 00:11:05.700 ]' 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1387 -- # jq '.[] .block_size' 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1387 -- # bs=512 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1388 -- # jq '.[] .num_blocks' 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1388 -- # nb=1048576 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1391 -- # bdev_size=512 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1392 -- # echo 512 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:11:05.700 19:07:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:07.615 19:07:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:11:07.615 19:07:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1202 -- # local i=0 00:11:07.615 19:07:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:11:07.615 19:07:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:11:07.615 19:07:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1209 -- # sleep 2 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1212 -- # return 0 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:11:09.531 19:07:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:11:09.792 19:07:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:11:10.053 19:07:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@76 -- # '[' 0 -eq 0 ']' 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@77 -- # run_test filesystem_ext4 nvmf_filesystem_create ext4 nvme0n1 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:10.996 ************************************ 00:11:10.996 START TEST filesystem_ext4 00:11:10.996 ************************************ 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create ext4 nvme0n1 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@930 -- # local fstype=ext4 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@932 -- # local i=0 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@933 -- # local force 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@935 -- # '[' ext4 = ext4 ']' 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@936 -- # force=-F 00:11:10.996 19:07:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@941 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:11:10.996 mke2fs 1.47.0 (5-Feb-2023) 00:11:10.996 Discarding device blocks: 0/522240 done 00:11:11.257 Creating filesystem with 522240 1k blocks and 130560 inodes 00:11:11.257 Filesystem UUID: e7c5c033-9447-4889-9bd9-5d2c2f9becce 00:11:11.257 Superblock backups stored on blocks: 00:11:11.257 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:11:11.257 00:11:11.257 Allocating group tables: 0/64 done 00:11:11.257 Writing inode tables: 0/64 done 00:11:12.200 Creating journal (8192 blocks): done 00:11:14.531 Writing superblocks and filesystem accounting information: 0/6450/64 done 00:11:14.531 00:11:14.531 19:07:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@949 -- # return 0 00:11:14.531 19:07:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@25 -- # sync 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@27 -- # sync 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@29 -- # i=0 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@37 -- # kill -0 3728597 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:11:21.126 00:11:21.126 real 0m9.043s 00:11:21.126 user 0m0.034s 00:11:21.126 sys 0m0.076s 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@10 -- # set +x 00:11:21.126 ************************************ 00:11:21.126 END TEST filesystem_ext4 00:11:21.126 ************************************ 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@78 -- # run_test filesystem_btrfs nvmf_filesystem_create btrfs nvme0n1 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:21.126 ************************************ 00:11:21.126 START TEST filesystem_btrfs 00:11:21.126 ************************************ 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create btrfs nvme0n1 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@930 -- # local fstype=btrfs 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@932 -- # local i=0 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@933 -- # local force 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@935 -- # '[' btrfs = ext4 ']' 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@938 -- # force=-f 00:11:21.126 19:07:32 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@941 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:11:21.126 btrfs-progs v6.8.1 00:11:21.127 See https://btrfs.readthedocs.io for more information. 00:11:21.127 00:11:21.127 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:11:21.127 NOTE: several default settings have changed in version 5.15, please make sure 00:11:21.127 this does not affect your deployments: 00:11:21.127 - DUP for metadata (-m dup) 00:11:21.127 - enabled no-holes (-O no-holes) 00:11:21.127 - enabled free-space-tree (-R free-space-tree) 00:11:21.127 00:11:21.127 Label: (null) 00:11:21.127 UUID: c4b21440-c082-4ab0-9514-79bfcf373c14 00:11:21.127 Node size: 16384 00:11:21.127 Sector size: 4096 (CPU page size: 4096) 00:11:21.127 Filesystem size: 510.00MiB 00:11:21.127 Block group profiles: 00:11:21.127 Data: single 8.00MiB 00:11:21.127 Metadata: DUP 32.00MiB 00:11:21.127 System: DUP 8.00MiB 00:11:21.127 SSD detected: yes 00:11:21.127 Zoned device: no 00:11:21.127 Features: extref, skinny-metadata, no-holes, free-space-tree 00:11:21.127 Checksum: crc32c 00:11:21.127 Number of devices: 1 00:11:21.127 Devices: 00:11:21.127 ID SIZE PATH 00:11:21.127 1 510.00MiB /dev/nvme0n1p1 00:11:21.127 00:11:21.127 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@949 -- # return 0 00:11:21.127 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:11:21.387 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:11:21.387 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@25 -- # sync 00:11:21.387 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:11:21.387 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@27 -- # sync 00:11:21.387 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@29 -- # i=0 00:11:21.387 19:07:33 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:11:21.387 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@37 -- # kill -0 3728597 00:11:21.387 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:11:21.387 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:11:21.647 00:11:21.647 real 0m1.361s 00:11:21.647 user 0m0.034s 00:11:21.647 sys 0m0.119s 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@10 -- # set +x 00:11:21.647 ************************************ 00:11:21.647 END TEST filesystem_btrfs 00:11:21.647 ************************************ 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@79 -- # run_test filesystem_xfs nvmf_filesystem_create xfs nvme0n1 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:21.647 ************************************ 00:11:21.647 START TEST filesystem_xfs 00:11:21.647 ************************************ 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create xfs nvme0n1 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@930 -- # local fstype=xfs 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@932 -- # local i=0 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@933 -- # local force 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@935 -- # '[' xfs = ext4 ']' 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@938 -- # force=-f 00:11:21.647 19:07:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@941 -- # mkfs.xfs -f /dev/nvme0n1p1 00:11:21.647 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:11:21.647 = sectsz=512 attr=2, projid32bit=1 00:11:21.647 = crc=1 finobt=1, sparse=1, rmapbt=0 00:11:21.647 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:11:21.647 data = bsize=4096 blocks=130560, imaxpct=25 00:11:21.647 = sunit=0 swidth=0 blks 00:11:21.647 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:11:21.647 log =internal log bsize=4096 blocks=16384, version=2 00:11:21.647 = sectsz=512 sunit=0 blks, lazy-count=1 00:11:21.647 realtime =none extsz=4096 blocks=0, rtextents=0 00:11:23.033 Discarding blocks...Done. 00:11:23.033 19:07:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@949 -- # return 0 00:11:23.033 19:07:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@25 -- # sync 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@27 -- # sync 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@29 -- # i=0 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@37 -- # kill -0 3728597 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:11:24.946 00:11:24.946 real 0m3.121s 00:11:24.946 user 0m0.023s 00:11:24.946 sys 0m0.084s 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@10 -- # set +x 00:11:24.946 ************************************ 00:11:24.946 END TEST filesystem_xfs 00:11:24.946 ************************************ 00:11:24.946 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@93 -- # sync 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:25.208 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1223 -- # local i=0 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1235 -- # return 0 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@101 -- # killprocess 3728597 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@954 -- # '[' -z 3728597 ']' 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@958 -- # kill -0 3728597 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@959 -- # uname 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3728597 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3728597' 00:11:25.208 killing process with pid 3728597 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@973 -- # kill 3728597 00:11:25.208 19:07:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@978 -- # wait 3728597 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:11:25.469 00:11:25.469 real 0m20.878s 00:11:25.469 user 1m22.572s 00:11:25.469 sys 0m1.431s 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:25.469 ************************************ 00:11:25.469 END TEST nvmf_filesystem_no_in_capsule 00:11:25.469 ************************************ 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@106 -- # run_test nvmf_filesystem_in_capsule nvmf_filesystem_part 4096 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:11:25.469 ************************************ 00:11:25.469 START TEST nvmf_filesystem_in_capsule 00:11:25.469 ************************************ 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1129 -- # nvmf_filesystem_part 4096 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@47 -- # in_capsule=4096 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:25.469 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@509 -- # nvmfpid=3732862 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@510 -- # waitforlisten 3732862 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@835 -- # '[' -z 3732862 ']' 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:25.730 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:25.730 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:25.730 [2024-11-26 19:07:38.149242] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:11:25.730 [2024-11-26 19:07:38.149297] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:25.730 [2024-11-26 19:07:38.236452] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:25.730 [2024-11-26 19:07:38.275817] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:25.730 [2024-11-26 19:07:38.275851] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:25.730 [2024-11-26 19:07:38.275858] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:25.730 [2024-11-26 19:07:38.275871] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:25.730 [2024-11-26 19:07:38.275877] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:25.730 [2024-11-26 19:07:38.277451] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:25.730 [2024-11-26 19:07:38.277571] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:25.730 [2024-11-26 19:07:38.277730] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:25.730 [2024-11-26 19:07:38.277732] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@868 -- # return 0 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 4096 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.674 19:07:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:26.674 [2024-11-26 19:07:39.005005] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:26.674 Malloc1 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:26.674 [2024-11-26 19:07:39.141902] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # local bdev_name=Malloc1 00:11:26.674 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # local bdev_info 00:11:26.675 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # local bs 00:11:26.675 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1385 -- # local nb 00:11:26.675 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1386 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:11:26.675 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.675 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:26.675 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.675 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1386 -- # bdev_info='[ 00:11:26.675 { 00:11:26.675 "name": "Malloc1", 00:11:26.675 "aliases": [ 00:11:26.675 "14b06eb1-5630-4d07-b661-477e896b260d" 00:11:26.675 ], 00:11:26.675 "product_name": "Malloc disk", 00:11:26.675 "block_size": 512, 00:11:26.675 "num_blocks": 1048576, 00:11:26.675 "uuid": "14b06eb1-5630-4d07-b661-477e896b260d", 00:11:26.675 "assigned_rate_limits": { 00:11:26.675 "rw_ios_per_sec": 0, 00:11:26.675 "rw_mbytes_per_sec": 0, 00:11:26.675 "r_mbytes_per_sec": 0, 00:11:26.675 "w_mbytes_per_sec": 0 00:11:26.675 }, 00:11:26.675 "claimed": true, 00:11:26.675 "claim_type": "exclusive_write", 00:11:26.675 "zoned": false, 00:11:26.675 "supported_io_types": { 00:11:26.675 "read": true, 00:11:26.675 "write": true, 00:11:26.675 "unmap": true, 00:11:26.675 "flush": true, 00:11:26.675 "reset": true, 00:11:26.675 "nvme_admin": false, 00:11:26.675 "nvme_io": false, 00:11:26.675 "nvme_io_md": false, 00:11:26.675 "write_zeroes": true, 00:11:26.675 "zcopy": true, 00:11:26.675 "get_zone_info": false, 00:11:26.675 "zone_management": false, 00:11:26.675 "zone_append": false, 00:11:26.675 "compare": false, 00:11:26.675 "compare_and_write": false, 00:11:26.675 "abort": true, 00:11:26.675 "seek_hole": false, 00:11:26.675 "seek_data": false, 00:11:26.675 "copy": true, 00:11:26.675 "nvme_iov_md": false 00:11:26.675 }, 00:11:26.675 "memory_domains": [ 00:11:26.675 { 00:11:26.676 "dma_device_id": "system", 00:11:26.676 "dma_device_type": 1 00:11:26.676 }, 00:11:26.676 { 00:11:26.676 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:26.676 "dma_device_type": 2 00:11:26.676 } 00:11:26.676 ], 00:11:26.676 "driver_specific": {} 00:11:26.676 } 00:11:26.676 ]' 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1387 -- # jq '.[] .block_size' 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1387 -- # bs=512 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1388 -- # jq '.[] .num_blocks' 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1388 -- # nb=1048576 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1391 -- # bdev_size=512 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1392 -- # echo 512 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:11:26.676 19:07:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:28.589 19:07:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:11:28.589 19:07:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1202 -- # local i=0 00:11:28.589 19:07:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:11:28.589 19:07:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:11:28.589 19:07:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1209 -- # sleep 2 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1212 -- # return 0 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:11:30.502 19:07:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:11:30.761 19:07:43 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:11:31.331 19:07:43 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@76 -- # '[' 4096 -eq 0 ']' 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@81 -- # run_test filesystem_in_capsule_ext4 nvmf_filesystem_create ext4 nvme0n1 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:32.275 ************************************ 00:11:32.275 START TEST filesystem_in_capsule_ext4 00:11:32.275 ************************************ 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create ext4 nvme0n1 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@930 -- # local fstype=ext4 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@932 -- # local i=0 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@933 -- # local force 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@935 -- # '[' ext4 = ext4 ']' 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@936 -- # force=-F 00:11:32.275 19:07:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@941 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:11:32.275 mke2fs 1.47.0 (5-Feb-2023) 00:11:32.275 Discarding device blocks: 0/522240 done 00:11:32.275 Creating filesystem with 522240 1k blocks and 130560 inodes 00:11:32.275 Filesystem UUID: c90a5707-11fd-438c-af29-efe60cf98723 00:11:32.275 Superblock backups stored on blocks: 00:11:32.275 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:11:32.275 00:11:32.275 Allocating group tables: 0/64 done 00:11:32.275 Writing inode tables: 0/64 done 00:11:33.218 Creating journal (8192 blocks): done 00:11:35.545 Writing superblocks and filesystem accounting information: 0/6450/64 done 00:11:35.545 00:11:35.545 19:07:48 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@949 -- # return 0 00:11:35.545 19:07:48 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@25 -- # sync 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@27 -- # sync 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@29 -- # i=0 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@37 -- # kill -0 3732862 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:11:42.125 00:11:42.125 real 0m9.592s 00:11:42.125 user 0m0.035s 00:11:42.125 sys 0m0.074s 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@10 -- # set +x 00:11:42.125 ************************************ 00:11:42.125 END TEST filesystem_in_capsule_ext4 00:11:42.125 ************************************ 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@82 -- # run_test filesystem_in_capsule_btrfs nvmf_filesystem_create btrfs nvme0n1 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:42.125 ************************************ 00:11:42.125 START TEST filesystem_in_capsule_btrfs 00:11:42.125 ************************************ 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create btrfs nvme0n1 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@930 -- # local fstype=btrfs 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@932 -- # local i=0 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@933 -- # local force 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@935 -- # '[' btrfs = ext4 ']' 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@938 -- # force=-f 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@941 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:11:42.125 btrfs-progs v6.8.1 00:11:42.125 See https://btrfs.readthedocs.io for more information. 00:11:42.125 00:11:42.125 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:11:42.125 NOTE: several default settings have changed in version 5.15, please make sure 00:11:42.125 this does not affect your deployments: 00:11:42.125 - DUP for metadata (-m dup) 00:11:42.125 - enabled no-holes (-O no-holes) 00:11:42.125 - enabled free-space-tree (-R free-space-tree) 00:11:42.125 00:11:42.125 Label: (null) 00:11:42.125 UUID: c5b6b0dc-bfbb-4e40-bfda-4f2d90348747 00:11:42.125 Node size: 16384 00:11:42.125 Sector size: 4096 (CPU page size: 4096) 00:11:42.125 Filesystem size: 510.00MiB 00:11:42.125 Block group profiles: 00:11:42.125 Data: single 8.00MiB 00:11:42.125 Metadata: DUP 32.00MiB 00:11:42.125 System: DUP 8.00MiB 00:11:42.125 SSD detected: yes 00:11:42.125 Zoned device: no 00:11:42.125 Features: extref, skinny-metadata, no-holes, free-space-tree 00:11:42.125 Checksum: crc32c 00:11:42.125 Number of devices: 1 00:11:42.125 Devices: 00:11:42.125 ID SIZE PATH 00:11:42.125 1 510.00MiB /dev/nvme0n1p1 00:11:42.125 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@949 -- # return 0 00:11:42.125 19:07:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@25 -- # sync 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@27 -- # sync 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@29 -- # i=0 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@37 -- # kill -0 3732862 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:11:43.067 00:11:43.067 real 0m1.148s 00:11:43.067 user 0m0.031s 00:11:43.067 sys 0m0.118s 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@10 -- # set +x 00:11:43.067 ************************************ 00:11:43.067 END TEST filesystem_in_capsule_btrfs 00:11:43.067 ************************************ 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@83 -- # run_test filesystem_in_capsule_xfs nvmf_filesystem_create xfs nvme0n1 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:43.067 ************************************ 00:11:43.067 START TEST filesystem_in_capsule_xfs 00:11:43.067 ************************************ 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create xfs nvme0n1 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@930 -- # local fstype=xfs 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@932 -- # local i=0 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@933 -- # local force 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@935 -- # '[' xfs = ext4 ']' 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@938 -- # force=-f 00:11:43.067 19:07:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@941 -- # mkfs.xfs -f /dev/nvme0n1p1 00:11:43.067 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:11:43.067 = sectsz=512 attr=2, projid32bit=1 00:11:43.067 = crc=1 finobt=1, sparse=1, rmapbt=0 00:11:43.067 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:11:43.067 data = bsize=4096 blocks=130560, imaxpct=25 00:11:43.067 = sunit=0 swidth=0 blks 00:11:43.067 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:11:43.067 log =internal log bsize=4096 blocks=16384, version=2 00:11:43.067 = sectsz=512 sunit=0 blks, lazy-count=1 00:11:43.067 realtime =none extsz=4096 blocks=0, rtextents=0 00:11:44.007 Discarding blocks...Done. 00:11:44.007 19:07:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@949 -- # return 0 00:11:44.007 19:07:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:11:45.917 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:11:45.917 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@25 -- # sync 00:11:45.917 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:11:45.917 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@27 -- # sync 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@29 -- # i=0 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@37 -- # kill -0 3732862 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:11:45.918 00:11:45.918 real 0m2.860s 00:11:45.918 user 0m0.024s 00:11:45.918 sys 0m0.081s 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@10 -- # set +x 00:11:45.918 ************************************ 00:11:45.918 END TEST filesystem_in_capsule_xfs 00:11:45.918 ************************************ 00:11:45.918 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@93 -- # sync 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:46.493 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1223 -- # local i=0 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1235 -- # return 0 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@101 -- # killprocess 3732862 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@954 -- # '[' -z 3732862 ']' 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@958 -- # kill -0 3732862 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@959 -- # uname 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:46.493 19:07:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3732862 00:11:46.493 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:46.494 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:46.494 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3732862' 00:11:46.494 killing process with pid 3732862 00:11:46.494 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@973 -- # kill 3732862 00:11:46.494 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@978 -- # wait 3732862 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:11:46.815 00:11:46.815 real 0m21.192s 00:11:46.815 user 1m23.802s 00:11:46.815 sys 0m1.497s 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:11:46.815 ************************************ 00:11:46.815 END TEST nvmf_filesystem_in_capsule 00:11:46.815 ************************************ 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@108 -- # nvmftestfini 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@516 -- # nvmfcleanup 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@121 -- # sync 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@124 -- # set +e 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:46.815 rmmod nvme_tcp 00:11:46.815 rmmod nvme_fabrics 00:11:46.815 rmmod nvme_keyring 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@128 -- # set -e 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@129 -- # return 0 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@297 -- # iptr 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@791 -- # iptables-save 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@791 -- # iptables-restore 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:46.815 19:07:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:49.455 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:49.455 00:11:49.455 real 0m53.191s 00:11:49.455 user 2m48.951s 00:11:49.455 sys 0m9.431s 00:11:49.455 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:11:49.456 ************************************ 00:11:49.456 END TEST nvmf_filesystem 00:11:49.456 ************************************ 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@18 -- # run_test nvmf_target_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:11:49.456 ************************************ 00:11:49.456 START TEST nvmf_target_discovery 00:11:49.456 ************************************ 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:11:49.456 * Looking for test storage... 00:11:49.456 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1693 -- # lcov --version 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@344 -- # case "$op" in 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@345 -- # : 1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # decimal 1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # decimal 2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # return 0 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:11:49.456 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:49.456 --rc genhtml_branch_coverage=1 00:11:49.456 --rc genhtml_function_coverage=1 00:11:49.456 --rc genhtml_legend=1 00:11:49.456 --rc geninfo_all_blocks=1 00:11:49.456 --rc geninfo_unexecuted_blocks=1 00:11:49.456 00:11:49.456 ' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:11:49.456 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:49.456 --rc genhtml_branch_coverage=1 00:11:49.456 --rc genhtml_function_coverage=1 00:11:49.456 --rc genhtml_legend=1 00:11:49.456 --rc geninfo_all_blocks=1 00:11:49.456 --rc geninfo_unexecuted_blocks=1 00:11:49.456 00:11:49.456 ' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:11:49.456 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:49.456 --rc genhtml_branch_coverage=1 00:11:49.456 --rc genhtml_function_coverage=1 00:11:49.456 --rc genhtml_legend=1 00:11:49.456 --rc geninfo_all_blocks=1 00:11:49.456 --rc geninfo_unexecuted_blocks=1 00:11:49.456 00:11:49.456 ' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:11:49.456 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:49.456 --rc genhtml_branch_coverage=1 00:11:49.456 --rc genhtml_function_coverage=1 00:11:49.456 --rc genhtml_legend=1 00:11:49.456 --rc geninfo_all_blocks=1 00:11:49.456 --rc geninfo_unexecuted_blocks=1 00:11:49.456 00:11:49.456 ' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # uname -s 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:49.456 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@5 -- # export PATH 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@51 -- # : 0 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:49.457 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@11 -- # NULL_BDEV_SIZE=102400 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@12 -- # NULL_BLOCK_SIZE=512 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@13 -- # NVMF_PORT_REFERRAL=4430 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@15 -- # hash nvme 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@20 -- # nvmftestinit 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@476 -- # prepare_net_devs 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@438 -- # local -g is_hw=no 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # remove_spdk_ns 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:11:49.457 19:08:01 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # e810=() 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # x722=() 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # mlx=() 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:57.601 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:11:57.602 Found 0000:31:00.0 (0x8086 - 0x159b) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:11:57.602 Found 0000:31:00.1 (0x8086 - 0x159b) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:11:57.602 Found net devices under 0000:31:00.0: cvl_0_0 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:11:57.602 Found net devices under 0000:31:00.1: cvl_0_1 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # is_hw=yes 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:57.602 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:57.602 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.681 ms 00:11:57.602 00:11:57.602 --- 10.0.0.2 ping statistics --- 00:11:57.602 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:57.602 rtt min/avg/max/mdev = 0.681/0.681/0.681/0.000 ms 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:57.602 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:57.602 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.291 ms 00:11:57.602 00:11:57.602 --- 10.0.0.1 ping statistics --- 00:11:57.602 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:57.602 rtt min/avg/max/mdev = 0.291/0.291/0.291/0.000 ms 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@450 -- # return 0 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:11:57.602 19:08:09 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@21 -- # nvmfappstart -m 0xF 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@509 -- # nvmfpid=3741835 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@510 -- # waitforlisten 3741835 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@835 -- # '[' -z 3741835 ']' 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:57.602 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:57.602 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:57.603 [2024-11-26 19:08:10.099197] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:11:57.603 [2024-11-26 19:08:10.099296] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:57.603 [2024-11-26 19:08:10.194312] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:57.864 [2024-11-26 19:08:10.236786] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:57.864 [2024-11-26 19:08:10.236824] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:57.864 [2024-11-26 19:08:10.236832] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:57.865 [2024-11-26 19:08:10.236839] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:57.865 [2024-11-26 19:08:10.236845] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:57.865 [2024-11-26 19:08:10.238494] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:57.865 [2024-11-26 19:08:10.238613] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:57.865 [2024-11-26 19:08:10.238772] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:57.865 [2024-11-26 19:08:10.238773] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@868 -- # return 0 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 [2024-11-26 19:08:10.959049] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # seq 1 4 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null1 102400 512 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 Null1 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Null1 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 [2024-11-26 19:08:11.019363] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null2 102400 512 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 Null2 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Null2 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.435 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null3 102400 512 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.696 Null3 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK00000000000003 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.696 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 Null3 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null4 102400 512 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 Null4 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode4 -a -s SPDK00000000000004 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode4 Null4 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode4 -t tcp -a 10.0.0.2 -s 4420 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@32 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@35 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 10.0.0.2 -s 4430 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.697 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@37 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 4420 00:11:58.959 00:11:58.959 Discovery Log Number of Records 6, Generation counter 6 00:11:58.959 =====Discovery Log Entry 0====== 00:11:58.959 trtype: tcp 00:11:58.959 adrfam: ipv4 00:11:58.959 subtype: current discovery subsystem 00:11:58.959 treq: not required 00:11:58.959 portid: 0 00:11:58.959 trsvcid: 4420 00:11:58.959 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:11:58.959 traddr: 10.0.0.2 00:11:58.959 eflags: explicit discovery connections, duplicate discovery information 00:11:58.959 sectype: none 00:11:58.959 =====Discovery Log Entry 1====== 00:11:58.959 trtype: tcp 00:11:58.959 adrfam: ipv4 00:11:58.959 subtype: nvme subsystem 00:11:58.959 treq: not required 00:11:58.959 portid: 0 00:11:58.959 trsvcid: 4420 00:11:58.959 subnqn: nqn.2016-06.io.spdk:cnode1 00:11:58.959 traddr: 10.0.0.2 00:11:58.959 eflags: none 00:11:58.959 sectype: none 00:11:58.959 =====Discovery Log Entry 2====== 00:11:58.959 trtype: tcp 00:11:58.959 adrfam: ipv4 00:11:58.959 subtype: nvme subsystem 00:11:58.959 treq: not required 00:11:58.959 portid: 0 00:11:58.959 trsvcid: 4420 00:11:58.959 subnqn: nqn.2016-06.io.spdk:cnode2 00:11:58.959 traddr: 10.0.0.2 00:11:58.959 eflags: none 00:11:58.959 sectype: none 00:11:58.959 =====Discovery Log Entry 3====== 00:11:58.959 trtype: tcp 00:11:58.959 adrfam: ipv4 00:11:58.959 subtype: nvme subsystem 00:11:58.959 treq: not required 00:11:58.959 portid: 0 00:11:58.959 trsvcid: 4420 00:11:58.959 subnqn: nqn.2016-06.io.spdk:cnode3 00:11:58.959 traddr: 10.0.0.2 00:11:58.959 eflags: none 00:11:58.959 sectype: none 00:11:58.959 =====Discovery Log Entry 4====== 00:11:58.959 trtype: tcp 00:11:58.959 adrfam: ipv4 00:11:58.959 subtype: nvme subsystem 00:11:58.959 treq: not required 00:11:58.959 portid: 0 00:11:58.959 trsvcid: 4420 00:11:58.959 subnqn: nqn.2016-06.io.spdk:cnode4 00:11:58.959 traddr: 10.0.0.2 00:11:58.959 eflags: none 00:11:58.959 sectype: none 00:11:58.959 =====Discovery Log Entry 5====== 00:11:58.959 trtype: tcp 00:11:58.959 adrfam: ipv4 00:11:58.959 subtype: discovery subsystem referral 00:11:58.959 treq: not required 00:11:58.959 portid: 0 00:11:58.959 trsvcid: 4430 00:11:58.959 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:11:58.959 traddr: 10.0.0.2 00:11:58.959 eflags: none 00:11:58.959 sectype: none 00:11:58.959 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@39 -- # echo 'Perform nvmf subsystem discovery via RPC' 00:11:58.959 Perform nvmf subsystem discovery via RPC 00:11:58.959 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@40 -- # rpc_cmd nvmf_get_subsystems 00:11:58.959 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.959 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.959 [ 00:11:58.959 { 00:11:58.959 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:11:58.959 "subtype": "Discovery", 00:11:58.959 "listen_addresses": [ 00:11:58.959 { 00:11:58.959 "trtype": "TCP", 00:11:58.959 "adrfam": "IPv4", 00:11:58.959 "traddr": "10.0.0.2", 00:11:58.959 "trsvcid": "4420" 00:11:58.959 } 00:11:58.959 ], 00:11:58.959 "allow_any_host": true, 00:11:58.959 "hosts": [] 00:11:58.959 }, 00:11:58.959 { 00:11:58.959 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:11:58.959 "subtype": "NVMe", 00:11:58.959 "listen_addresses": [ 00:11:58.959 { 00:11:58.959 "trtype": "TCP", 00:11:58.959 "adrfam": "IPv4", 00:11:58.959 "traddr": "10.0.0.2", 00:11:58.959 "trsvcid": "4420" 00:11:58.959 } 00:11:58.959 ], 00:11:58.959 "allow_any_host": true, 00:11:58.959 "hosts": [], 00:11:58.959 "serial_number": "SPDK00000000000001", 00:11:58.959 "model_number": "SPDK bdev Controller", 00:11:58.959 "max_namespaces": 32, 00:11:58.959 "min_cntlid": 1, 00:11:58.959 "max_cntlid": 65519, 00:11:58.959 "namespaces": [ 00:11:58.959 { 00:11:58.959 "nsid": 1, 00:11:58.959 "bdev_name": "Null1", 00:11:58.959 "name": "Null1", 00:11:58.959 "nguid": "69E3E6CDF7704076B47CBB485D47641C", 00:11:58.959 "uuid": "69e3e6cd-f770-4076-b47c-bb485d47641c" 00:11:58.959 } 00:11:58.959 ] 00:11:58.959 }, 00:11:58.959 { 00:11:58.959 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:11:58.960 "subtype": "NVMe", 00:11:58.960 "listen_addresses": [ 00:11:58.960 { 00:11:58.960 "trtype": "TCP", 00:11:58.960 "adrfam": "IPv4", 00:11:58.960 "traddr": "10.0.0.2", 00:11:58.960 "trsvcid": "4420" 00:11:58.960 } 00:11:58.960 ], 00:11:58.960 "allow_any_host": true, 00:11:58.960 "hosts": [], 00:11:58.960 "serial_number": "SPDK00000000000002", 00:11:58.960 "model_number": "SPDK bdev Controller", 00:11:58.960 "max_namespaces": 32, 00:11:58.960 "min_cntlid": 1, 00:11:58.960 "max_cntlid": 65519, 00:11:58.960 "namespaces": [ 00:11:58.960 { 00:11:58.960 "nsid": 1, 00:11:58.960 "bdev_name": "Null2", 00:11:58.960 "name": "Null2", 00:11:58.960 "nguid": "DB4C2AA0BF224D7DA6CBCD68B1285681", 00:11:58.960 "uuid": "db4c2aa0-bf22-4d7d-a6cb-cd68b1285681" 00:11:58.960 } 00:11:58.960 ] 00:11:58.960 }, 00:11:58.960 { 00:11:58.960 "nqn": "nqn.2016-06.io.spdk:cnode3", 00:11:58.960 "subtype": "NVMe", 00:11:58.960 "listen_addresses": [ 00:11:58.960 { 00:11:58.960 "trtype": "TCP", 00:11:58.960 "adrfam": "IPv4", 00:11:58.960 "traddr": "10.0.0.2", 00:11:58.960 "trsvcid": "4420" 00:11:58.960 } 00:11:58.960 ], 00:11:58.960 "allow_any_host": true, 00:11:58.960 "hosts": [], 00:11:58.960 "serial_number": "SPDK00000000000003", 00:11:58.960 "model_number": "SPDK bdev Controller", 00:11:58.960 "max_namespaces": 32, 00:11:58.960 "min_cntlid": 1, 00:11:58.960 "max_cntlid": 65519, 00:11:58.960 "namespaces": [ 00:11:58.960 { 00:11:58.960 "nsid": 1, 00:11:58.960 "bdev_name": "Null3", 00:11:58.960 "name": "Null3", 00:11:58.960 "nguid": "D79C42DFD15448DCA01A721D3D4DC275", 00:11:58.960 "uuid": "d79c42df-d154-48dc-a01a-721d3d4dc275" 00:11:58.960 } 00:11:58.960 ] 00:11:58.960 }, 00:11:58.960 { 00:11:58.960 "nqn": "nqn.2016-06.io.spdk:cnode4", 00:11:58.960 "subtype": "NVMe", 00:11:58.960 "listen_addresses": [ 00:11:58.960 { 00:11:58.960 "trtype": "TCP", 00:11:58.960 "adrfam": "IPv4", 00:11:58.960 "traddr": "10.0.0.2", 00:11:58.960 "trsvcid": "4420" 00:11:58.960 } 00:11:58.960 ], 00:11:58.960 "allow_any_host": true, 00:11:58.960 "hosts": [], 00:11:58.960 "serial_number": "SPDK00000000000004", 00:11:58.960 "model_number": "SPDK bdev Controller", 00:11:58.960 "max_namespaces": 32, 00:11:58.960 "min_cntlid": 1, 00:11:58.960 "max_cntlid": 65519, 00:11:58.960 "namespaces": [ 00:11:58.960 { 00:11:58.960 "nsid": 1, 00:11:58.960 "bdev_name": "Null4", 00:11:58.960 "name": "Null4", 00:11:58.960 "nguid": "15C3B1B488664B23BCDAC8919B4429E0", 00:11:58.960 "uuid": "15c3b1b4-8866-4b23-bcda-c8919b4429e0" 00:11:58.960 } 00:11:58.960 ] 00:11:58.960 } 00:11:58.960 ] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # seq 1 4 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null1 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null2 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null3 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode4 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null4 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@47 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 10.0.0.2 -s 4430 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # rpc_cmd bdev_get_bdevs 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # jq -r '.[].name' 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # check_bdevs= 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@50 -- # '[' -n '' ']' 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@55 -- # trap - SIGINT SIGTERM EXIT 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@57 -- # nvmftestfini 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@516 -- # nvmfcleanup 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@121 -- # sync 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@124 -- # set +e 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:58.960 rmmod nvme_tcp 00:11:58.960 rmmod nvme_fabrics 00:11:58.960 rmmod nvme_keyring 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@128 -- # set -e 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@129 -- # return 0 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@517 -- # '[' -n 3741835 ']' 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@518 -- # killprocess 3741835 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@954 -- # '[' -z 3741835 ']' 00:11:58.960 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@958 -- # kill -0 3741835 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@959 -- # uname 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3741835 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3741835' 00:11:59.221 killing process with pid 3741835 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@973 -- # kill 3741835 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@978 -- # wait 3741835 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@297 -- # iptr 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@791 -- # iptables-save 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@791 -- # iptables-restore 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:59.221 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:59.222 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:59.222 19:08:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:01.765 00:12:01.765 real 0m12.308s 00:12:01.765 user 0m8.734s 00:12:01.765 sys 0m6.594s 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:12:01.765 ************************************ 00:12:01.765 END TEST nvmf_target_discovery 00:12:01.765 ************************************ 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@19 -- # run_test nvmf_referrals /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:12:01.765 ************************************ 00:12:01.765 START TEST nvmf_referrals 00:12:01.765 ************************************ 00:12:01.765 19:08:13 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:12:01.765 * Looking for test storage... 00:12:01.765 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1693 -- # lcov --version 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # IFS=.-: 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # read -ra ver1 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # IFS=.-: 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # read -ra ver2 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@338 -- # local 'op=<' 00:12:01.765 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@340 -- # ver1_l=2 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@341 -- # ver2_l=1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@344 -- # case "$op" in 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@345 -- # : 1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # decimal 1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # ver1[v]=1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # decimal 2 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=2 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 2 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # ver2[v]=2 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # return 0 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:12:01.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:01.766 --rc genhtml_branch_coverage=1 00:12:01.766 --rc genhtml_function_coverage=1 00:12:01.766 --rc genhtml_legend=1 00:12:01.766 --rc geninfo_all_blocks=1 00:12:01.766 --rc geninfo_unexecuted_blocks=1 00:12:01.766 00:12:01.766 ' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:12:01.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:01.766 --rc genhtml_branch_coverage=1 00:12:01.766 --rc genhtml_function_coverage=1 00:12:01.766 --rc genhtml_legend=1 00:12:01.766 --rc geninfo_all_blocks=1 00:12:01.766 --rc geninfo_unexecuted_blocks=1 00:12:01.766 00:12:01.766 ' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:12:01.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:01.766 --rc genhtml_branch_coverage=1 00:12:01.766 --rc genhtml_function_coverage=1 00:12:01.766 --rc genhtml_legend=1 00:12:01.766 --rc geninfo_all_blocks=1 00:12:01.766 --rc geninfo_unexecuted_blocks=1 00:12:01.766 00:12:01.766 ' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:12:01.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:01.766 --rc genhtml_branch_coverage=1 00:12:01.766 --rc genhtml_function_coverage=1 00:12:01.766 --rc genhtml_legend=1 00:12:01.766 --rc geninfo_all_blocks=1 00:12:01.766 --rc geninfo_unexecuted_blocks=1 00:12:01.766 00:12:01.766 ' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # uname -s 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@15 -- # shopt -s extglob 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@5 -- # export PATH 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@51 -- # : 0 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:01.766 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@11 -- # NVMF_REFERRAL_IP_1=127.0.0.2 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@12 -- # NVMF_REFERRAL_IP_2=127.0.0.3 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@13 -- # NVMF_REFERRAL_IP_3=127.0.0.4 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@14 -- # NVMF_PORT_REFERRAL=4430 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@15 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@16 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@37 -- # nvmftestinit 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:01.766 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:01.767 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:01.767 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:01.767 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@309 -- # xtrace_disable 00:12:01.767 19:08:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # pci_devs=() 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # net_devs=() 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # e810=() 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # local -ga e810 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # x722=() 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # local -ga x722 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # mlx=() 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # local -ga mlx 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:12:09.907 Found 0000:31:00.0 (0x8086 - 0x159b) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:12:09.907 Found 0000:31:00.1 (0x8086 - 0x159b) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:12:09.907 Found net devices under 0000:31:00.0: cvl_0_0 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:12:09.907 Found net devices under 0000:31:00.1: cvl_0_1 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # is_hw=yes 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:09.907 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:09.908 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:09.908 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.702 ms 00:12:09.908 00:12:09.908 --- 10.0.0.2 ping statistics --- 00:12:09.908 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:09.908 rtt min/avg/max/mdev = 0.702/0.702/0.702/0.000 ms 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:09.908 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:09.908 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.283 ms 00:12:09.908 00:12:09.908 --- 10.0.0.1 ping statistics --- 00:12:09.908 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:09.908 rtt min/avg/max/mdev = 0.283/0.283/0.283/0.000 ms 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@450 -- # return 0 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:12:09.908 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@38 -- # nvmfappstart -m 0xF 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@726 -- # xtrace_disable 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@509 -- # nvmfpid=3747028 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@510 -- # waitforlisten 3747028 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@835 -- # '[' -z 3747028 ']' 00:12:10.168 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:10.169 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:10.169 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:10.169 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:10.169 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:10.169 19:08:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:10.169 [2024-11-26 19:08:22.616736] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:12:10.169 [2024-11-26 19:08:22.616806] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:10.169 [2024-11-26 19:08:22.708770] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:10.169 [2024-11-26 19:08:22.750775] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:10.169 [2024-11-26 19:08:22.750816] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:10.169 [2024-11-26 19:08:22.750824] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:10.169 [2024-11-26 19:08:22.750831] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:10.169 [2024-11-26 19:08:22.750837] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:10.169 [2024-11-26 19:08:22.752715] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:10.169 [2024-11-26 19:08:22.752851] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:12:10.169 [2024-11-26 19:08:22.753019] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:10.169 [2024-11-26 19:08:22.753019] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@868 -- # return 0 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@732 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@40 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 [2024-11-26 19:08:23.476936] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 10.0.0.2 -s 8009 discovery 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 [2024-11-26 19:08:23.504013] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@44 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@45 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.3 -s 4430 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@46 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.4 -s 4430 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # rpc_cmd nvmf_discovery_get_referrals 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # jq length 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # (( 3 == 3 )) 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # get_referral_ips rpc 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # get_referral_ips nvme 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:12:11.110 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@52 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@53 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.3 -s 4430 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@54 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.4 -s 4430 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # rpc_cmd nvmf_discovery_get_referrals 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # jq length 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # (( 0 == 0 )) 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # get_referral_ips nvme 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:12:11.372 19:08:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # [[ '' == '' ]] 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@60 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n discovery 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@62 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:12:11.632 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # get_referral_ips rpc 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.2 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # get_referral_ips nvme 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:12:11.633 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:12:11.893 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.2 00:12:11.893 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:12:11.893 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # get_discovery_entries 'nvme subsystem' 00:12:11.893 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # jq -r .subnqn 00:12:11.893 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:12:11.893 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:11.894 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:12:11.894 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # [[ nqn.2016-06.io.spdk:cnode1 == \n\q\n\.\2\0\1\6\-\0\6\.\i\o\.\s\p\d\k\:\c\n\o\d\e\1 ]] 00:12:11.894 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # get_discovery_entries 'discovery subsystem referral' 00:12:11.894 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # jq -r .subnqn 00:12:11.894 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:12:11.894 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:11.894 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@71 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # get_referral_ips rpc 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # get_referral_ips nvme 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:12:12.155 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:12:12.415 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 00:12:12.415 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:12:12.416 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # get_discovery_entries 'nvme subsystem' 00:12:12.416 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # jq -r .subnqn 00:12:12.416 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:12:12.416 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:12.416 19:08:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # [[ '' == '' ]] 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # get_discovery_entries 'discovery subsystem referral' 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # jq -r .subnqn 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@79 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2014-08.org.nvmexpress.discovery 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # rpc_cmd nvmf_discovery_get_referrals 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # jq length 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:12.676 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # (( 0 == 0 )) 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # get_referral_ips nvme 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 8009 -o json 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # [[ '' == '' ]] 00:12:12.936 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@85 -- # trap - SIGINT SIGTERM EXIT 00:12:12.937 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@86 -- # nvmftestfini 00:12:12.937 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:12.937 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@121 -- # sync 00:12:12.937 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:12.937 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@124 -- # set +e 00:12:12.937 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:12.937 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:12.937 rmmod nvme_tcp 00:12:13.197 rmmod nvme_fabrics 00:12:13.197 rmmod nvme_keyring 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@128 -- # set -e 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@129 -- # return 0 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@517 -- # '[' -n 3747028 ']' 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@518 -- # killprocess 3747028 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@954 -- # '[' -z 3747028 ']' 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@958 -- # kill -0 3747028 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@959 -- # uname 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3747028 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3747028' 00:12:13.197 killing process with pid 3747028 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@973 -- # kill 3747028 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@978 -- # wait 3747028 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@297 -- # iptr 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@791 -- # iptables-restore 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@791 -- # iptables-save 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:13.197 19:08:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:15.742 00:12:15.742 real 0m13.944s 00:12:15.742 user 0m15.587s 00:12:15.742 sys 0m7.053s 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:12:15.742 ************************************ 00:12:15.742 END TEST nvmf_referrals 00:12:15.742 ************************************ 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@20 -- # run_test nvmf_connect_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:12:15.742 ************************************ 00:12:15.742 START TEST nvmf_connect_disconnect 00:12:15.742 ************************************ 00:12:15.742 19:08:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:12:15.742 * Looking for test storage... 00:12:15.742 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1693 -- # lcov --version 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@345 -- # : 1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # decimal 1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # decimal 2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # return 0 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:12:15.743 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:15.743 --rc genhtml_branch_coverage=1 00:12:15.743 --rc genhtml_function_coverage=1 00:12:15.743 --rc genhtml_legend=1 00:12:15.743 --rc geninfo_all_blocks=1 00:12:15.743 --rc geninfo_unexecuted_blocks=1 00:12:15.743 00:12:15.743 ' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:12:15.743 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:15.743 --rc genhtml_branch_coverage=1 00:12:15.743 --rc genhtml_function_coverage=1 00:12:15.743 --rc genhtml_legend=1 00:12:15.743 --rc geninfo_all_blocks=1 00:12:15.743 --rc geninfo_unexecuted_blocks=1 00:12:15.743 00:12:15.743 ' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:12:15.743 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:15.743 --rc genhtml_branch_coverage=1 00:12:15.743 --rc genhtml_function_coverage=1 00:12:15.743 --rc genhtml_legend=1 00:12:15.743 --rc geninfo_all_blocks=1 00:12:15.743 --rc geninfo_unexecuted_blocks=1 00:12:15.743 00:12:15.743 ' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:12:15.743 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:15.743 --rc genhtml_branch_coverage=1 00:12:15.743 --rc genhtml_function_coverage=1 00:12:15.743 --rc genhtml_legend=1 00:12:15.743 --rc geninfo_all_blocks=1 00:12:15.743 --rc geninfo_unexecuted_blocks=1 00:12:15.743 00:12:15.743 ' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # uname -s 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@5 -- # export PATH 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@51 -- # : 0 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:15.743 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:15.744 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@11 -- # MALLOC_BDEV_SIZE=64 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@15 -- # nvmftestinit 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:12:15.744 19:08:28 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # e810=() 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # x722=() 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:12:23.939 Found 0000:31:00.0 (0x8086 - 0x159b) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:12:23.939 Found 0000:31:00.1 (0x8086 - 0x159b) 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:23.939 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:12:23.940 Found net devices under 0000:31:00.0: cvl_0_0 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:12:23.940 Found net devices under 0000:31:00.1: cvl_0_1 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # is_hw=yes 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:23.940 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:23.940 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.656 ms 00:12:23.940 00:12:23.940 --- 10.0.0.2 ping statistics --- 00:12:23.940 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:23.940 rtt min/avg/max/mdev = 0.656/0.656/0.656/0.000 ms 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:23.940 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:23.940 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.303 ms 00:12:23.940 00:12:23.940 --- 10.0.0.1 ping statistics --- 00:12:23.940 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:23.940 rtt min/avg/max/mdev = 0.303/0.303/0.303/0.000 ms 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@450 -- # return 0 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:12:23.940 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@16 -- # nvmfappstart -m 0xF 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@726 -- # xtrace_disable 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@509 -- # nvmfpid=3752469 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@510 -- # waitforlisten 3752469 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@835 -- # '[' -z 3752469 ']' 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:24.201 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:24.201 19:08:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:24.201 [2024-11-26 19:08:36.652256] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:12:24.201 [2024-11-26 19:08:36.652326] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:24.201 [2024-11-26 19:08:36.746425] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:24.201 [2024-11-26 19:08:36.785805] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:24.201 [2024-11-26 19:08:36.785842] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:24.201 [2024-11-26 19:08:36.785851] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:24.201 [2024-11-26 19:08:36.785857] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:24.201 [2024-11-26 19:08:36.785869] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:24.201 [2024-11-26 19:08:36.787389] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:24.201 [2024-11-26 19:08:36.787477] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:12:24.201 [2024-11-26 19:08:36.787638] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:24.201 [2024-11-26 19:08:36.787639] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@868 -- # return 0 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@732 -- # xtrace_disable 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:25.145 [2024-11-26 19:08:37.499125] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # bdev=Malloc0 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:25.145 [2024-11-26 19:08:37.570306] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@26 -- # '[' 0 -eq 1 ']' 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@31 -- # num_iterations=5 00:12:25.145 19:08:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@34 -- # set +x 00:12:29.346 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:32.641 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:36.034 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:40.234 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:43.530 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@43 -- # trap - SIGINT SIGTERM EXIT 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@45 -- # nvmftestfini 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@121 -- # sync 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@124 -- # set +e 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:43.530 rmmod nvme_tcp 00:12:43.530 rmmod nvme_fabrics 00:12:43.530 rmmod nvme_keyring 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@128 -- # set -e 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@129 -- # return 0 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@517 -- # '[' -n 3752469 ']' 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@518 -- # killprocess 3752469 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@954 -- # '[' -z 3752469 ']' 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@958 -- # kill -0 3752469 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@959 -- # uname 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3752469 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3752469' 00:12:43.530 killing process with pid 3752469 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@973 -- # kill 3752469 00:12:43.530 19:08:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@978 -- # wait 3752469 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@297 -- # iptr 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@791 -- # iptables-save 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@791 -- # iptables-restore 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:43.530 19:08:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:46.072 00:12:46.072 real 0m30.165s 00:12:46.072 user 1m19.074s 00:12:46.072 sys 0m7.787s 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:12:46.072 ************************************ 00:12:46.072 END TEST nvmf_connect_disconnect 00:12:46.072 ************************************ 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@21 -- # run_test nvmf_multitarget /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:12:46.072 ************************************ 00:12:46.072 START TEST nvmf_multitarget 00:12:46.072 ************************************ 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:12:46.072 * Looking for test storage... 00:12:46.072 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1693 -- # lcov --version 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # IFS=.-: 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # read -ra ver1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # IFS=.-: 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # read -ra ver2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@338 -- # local 'op=<' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@340 -- # ver1_l=2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@341 -- # ver2_l=1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@344 -- # case "$op" in 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@345 -- # : 1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # decimal 1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # ver1[v]=1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # decimal 2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # ver2[v]=2 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # return 0 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:12:46.072 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:46.072 --rc genhtml_branch_coverage=1 00:12:46.072 --rc genhtml_function_coverage=1 00:12:46.072 --rc genhtml_legend=1 00:12:46.072 --rc geninfo_all_blocks=1 00:12:46.072 --rc geninfo_unexecuted_blocks=1 00:12:46.072 00:12:46.072 ' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:12:46.072 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:46.072 --rc genhtml_branch_coverage=1 00:12:46.072 --rc genhtml_function_coverage=1 00:12:46.072 --rc genhtml_legend=1 00:12:46.072 --rc geninfo_all_blocks=1 00:12:46.072 --rc geninfo_unexecuted_blocks=1 00:12:46.072 00:12:46.072 ' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:12:46.072 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:46.072 --rc genhtml_branch_coverage=1 00:12:46.072 --rc genhtml_function_coverage=1 00:12:46.072 --rc genhtml_legend=1 00:12:46.072 --rc geninfo_all_blocks=1 00:12:46.072 --rc geninfo_unexecuted_blocks=1 00:12:46.072 00:12:46.072 ' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:12:46.072 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:46.072 --rc genhtml_branch_coverage=1 00:12:46.072 --rc genhtml_function_coverage=1 00:12:46.072 --rc genhtml_legend=1 00:12:46.072 --rc geninfo_all_blocks=1 00:12:46.072 --rc geninfo_unexecuted_blocks=1 00:12:46.072 00:12:46.072 ' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # uname -s 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@15 -- # shopt -s extglob 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:46.072 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@5 -- # export PATH 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@51 -- # : 0 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:46.073 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@13 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@15 -- # nvmftestinit 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@309 -- # xtrace_disable 00:12:46.073 19:08:58 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # pci_devs=() 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # net_devs=() 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # e810=() 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # local -ga e810 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # x722=() 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # local -ga x722 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # mlx=() 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # local -ga mlx 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:54.202 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:12:54.202 Found 0000:31:00.0 (0x8086 - 0x159b) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:12:54.203 Found 0000:31:00.1 (0x8086 - 0x159b) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:12:54.203 Found net devices under 0000:31:00.0: cvl_0_0 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:12:54.203 Found net devices under 0000:31:00.1: cvl_0_1 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # is_hw=yes 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:54.203 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:54.203 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.669 ms 00:12:54.203 00:12:54.203 --- 10.0.0.2 ping statistics --- 00:12:54.203 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:54.203 rtt min/avg/max/mdev = 0.669/0.669/0.669/0.000 ms 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:54.203 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:54.203 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.329 ms 00:12:54.203 00:12:54.203 --- 10.0.0.1 ping statistics --- 00:12:54.203 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:54.203 rtt min/avg/max/mdev = 0.329/0.329/0.329/0.000 ms 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@450 -- # return 0 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:12:54.203 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@16 -- # nvmfappstart -m 0xF 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@726 -- # xtrace_disable 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@509 -- # nvmfpid=3761220 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@510 -- # waitforlisten 3761220 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@835 -- # '[' -z 3761220 ']' 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:54.463 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:54.463 19:09:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:54.463 [2024-11-26 19:09:06.941929] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:12:54.463 [2024-11-26 19:09:06.941998] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:54.463 [2024-11-26 19:09:07.033987] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:54.463 [2024-11-26 19:09:07.074984] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:54.463 [2024-11-26 19:09:07.075016] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:54.463 [2024-11-26 19:09:07.075024] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:54.463 [2024-11-26 19:09:07.075031] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:54.463 [2024-11-26 19:09:07.075036] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:54.463 [2024-11-26 19:09:07.076798] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:54.463 [2024-11-26 19:09:07.076934] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:12:54.463 [2024-11-26 19:09:07.076991] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:54.463 [2024-11-26 19:09:07.076992] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@868 -- # return 0 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@732 -- # xtrace_disable 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@18 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # jq length 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # '[' 1 '!=' 1 ']' 00:12:55.404 19:09:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_1 -s 32 00:12:55.404 "nvmf_tgt_1" 00:12:55.404 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_2 -s 32 00:12:55.664 "nvmf_tgt_2" 00:12:55.664 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:12:55.664 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # jq length 00:12:55.664 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # '[' 3 '!=' 3 ']' 00:12:55.664 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_1 00:12:55.925 true 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_2 00:12:55.925 true 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # jq length 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # '[' 1 '!=' 1 ']' 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@41 -- # nvmftestfini 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@121 -- # sync 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@124 -- # set +e 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:55.925 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:56.186 rmmod nvme_tcp 00:12:56.186 rmmod nvme_fabrics 00:12:56.186 rmmod nvme_keyring 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@128 -- # set -e 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@129 -- # return 0 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@517 -- # '[' -n 3761220 ']' 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@518 -- # killprocess 3761220 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@954 -- # '[' -z 3761220 ']' 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@958 -- # kill -0 3761220 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@959 -- # uname 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3761220 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3761220' 00:12:56.186 killing process with pid 3761220 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@973 -- # kill 3761220 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@978 -- # wait 3761220 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@297 -- # iptr 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@791 -- # iptables-save 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@791 -- # iptables-restore 00:12:56.186 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:56.447 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:56.447 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:56.447 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:56.447 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:56.447 19:09:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:58.356 00:12:58.356 real 0m12.686s 00:12:58.356 user 0m10.148s 00:12:58.356 sys 0m6.811s 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:58.356 ************************************ 00:12:58.356 END TEST nvmf_multitarget 00:12:58.356 ************************************ 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@22 -- # run_test nvmf_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:12:58.356 ************************************ 00:12:58.356 START TEST nvmf_rpc 00:12:58.356 ************************************ 00:12:58.356 19:09:10 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:12:58.618 * Looking for test storage... 00:12:58.618 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@344 -- # case "$op" in 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@345 -- # : 1 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # decimal 1 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=1 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 1 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # decimal 2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # return 0 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:12:58.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:58.618 --rc genhtml_branch_coverage=1 00:12:58.618 --rc genhtml_function_coverage=1 00:12:58.618 --rc genhtml_legend=1 00:12:58.618 --rc geninfo_all_blocks=1 00:12:58.618 --rc geninfo_unexecuted_blocks=1 00:12:58.618 00:12:58.618 ' 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:12:58.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:58.618 --rc genhtml_branch_coverage=1 00:12:58.618 --rc genhtml_function_coverage=1 00:12:58.618 --rc genhtml_legend=1 00:12:58.618 --rc geninfo_all_blocks=1 00:12:58.618 --rc geninfo_unexecuted_blocks=1 00:12:58.618 00:12:58.618 ' 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:12:58.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:58.618 --rc genhtml_branch_coverage=1 00:12:58.618 --rc genhtml_function_coverage=1 00:12:58.618 --rc genhtml_legend=1 00:12:58.618 --rc geninfo_all_blocks=1 00:12:58.618 --rc geninfo_unexecuted_blocks=1 00:12:58.618 00:12:58.618 ' 00:12:58.618 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:12:58.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:58.618 --rc genhtml_branch_coverage=1 00:12:58.619 --rc genhtml_function_coverage=1 00:12:58.619 --rc genhtml_legend=1 00:12:58.619 --rc geninfo_all_blocks=1 00:12:58.619 --rc geninfo_unexecuted_blocks=1 00:12:58.619 00:12:58.619 ' 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # uname -s 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@15 -- # shopt -s extglob 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@5 -- # export PATH 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@51 -- # : 0 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:58.619 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@11 -- # loops=5 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@23 -- # nvmftestinit 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@309 -- # xtrace_disable 00:12:58.619 19:09:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # pci_devs=() 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # net_devs=() 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # e810=() 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # local -ga e810 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # x722=() 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # local -ga x722 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # mlx=() 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # local -ga mlx 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:06.767 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:13:06.768 Found 0000:31:00.0 (0x8086 - 0x159b) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:13:06.768 Found 0000:31:00.1 (0x8086 - 0x159b) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:13:06.768 Found net devices under 0000:31:00.0: cvl_0_0 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:13:06.768 Found net devices under 0000:31:00.1: cvl_0_1 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # is_hw=yes 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:13:06.768 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:06.769 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:07.029 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:07.029 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.576 ms 00:13:07.029 00:13:07.029 --- 10.0.0.2 ping statistics --- 00:13:07.029 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:07.029 rtt min/avg/max/mdev = 0.576/0.576/0.576/0.000 ms 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:07.029 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:07.029 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.282 ms 00:13:07.029 00:13:07.029 --- 10.0.0.1 ping statistics --- 00:13:07.029 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:07.029 rtt min/avg/max/mdev = 0.282/0.282/0.282/0.000 ms 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@450 -- # return 0 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@24 -- # nvmfappstart -m 0xF 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@726 -- # xtrace_disable 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@509 -- # nvmfpid=3766725 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@510 -- # waitforlisten 3766725 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@835 -- # '[' -z 3766725 ']' 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:07.029 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:07.029 19:09:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:07.289 [2024-11-26 19:09:19.664308] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:13:07.289 [2024-11-26 19:09:19.664361] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:07.289 [2024-11-26 19:09:19.750671] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:07.289 [2024-11-26 19:09:19.788542] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:07.289 [2024-11-26 19:09:19.788574] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:07.289 [2024-11-26 19:09:19.788582] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:07.289 [2024-11-26 19:09:19.788589] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:07.289 [2024-11-26 19:09:19.788599] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:07.289 [2024-11-26 19:09:19.790092] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:07.289 [2024-11-26 19:09:19.790208] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:07.289 [2024-11-26 19:09:19.790365] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:07.289 [2024-11-26 19:09:19.790366] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:07.858 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:07.858 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@868 -- # return 0 00:13:07.858 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:13:07.858 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@732 -- # xtrace_disable 00:13:07.858 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # rpc_cmd nvmf_get_stats 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # stats='{ 00:13:08.119 "tick_rate": 2400000000, 00:13:08.119 "poll_groups": [ 00:13:08.119 { 00:13:08.119 "name": "nvmf_tgt_poll_group_000", 00:13:08.119 "admin_qpairs": 0, 00:13:08.119 "io_qpairs": 0, 00:13:08.119 "current_admin_qpairs": 0, 00:13:08.119 "current_io_qpairs": 0, 00:13:08.119 "pending_bdev_io": 0, 00:13:08.119 "completed_nvme_io": 0, 00:13:08.119 "transports": [] 00:13:08.119 }, 00:13:08.119 { 00:13:08.119 "name": "nvmf_tgt_poll_group_001", 00:13:08.119 "admin_qpairs": 0, 00:13:08.119 "io_qpairs": 0, 00:13:08.119 "current_admin_qpairs": 0, 00:13:08.119 "current_io_qpairs": 0, 00:13:08.119 "pending_bdev_io": 0, 00:13:08.119 "completed_nvme_io": 0, 00:13:08.119 "transports": [] 00:13:08.119 }, 00:13:08.119 { 00:13:08.119 "name": "nvmf_tgt_poll_group_002", 00:13:08.119 "admin_qpairs": 0, 00:13:08.119 "io_qpairs": 0, 00:13:08.119 "current_admin_qpairs": 0, 00:13:08.119 "current_io_qpairs": 0, 00:13:08.119 "pending_bdev_io": 0, 00:13:08.119 "completed_nvme_io": 0, 00:13:08.119 "transports": [] 00:13:08.119 }, 00:13:08.119 { 00:13:08.119 "name": "nvmf_tgt_poll_group_003", 00:13:08.119 "admin_qpairs": 0, 00:13:08.119 "io_qpairs": 0, 00:13:08.119 "current_admin_qpairs": 0, 00:13:08.119 "current_io_qpairs": 0, 00:13:08.119 "pending_bdev_io": 0, 00:13:08.119 "completed_nvme_io": 0, 00:13:08.119 "transports": [] 00:13:08.119 } 00:13:08.119 ] 00:13:08.119 }' 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # jcount '.poll_groups[].name' 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@14 -- # local 'filter=.poll_groups[].name' 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # jq '.poll_groups[].name' 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # wc -l 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # (( 4 == 4 )) 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # jq '.poll_groups[0].transports[0]' 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # [[ null == null ]] 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@31 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.119 [2024-11-26 19:09:20.633729] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # rpc_cmd nvmf_get_stats 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.119 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # stats='{ 00:13:08.119 "tick_rate": 2400000000, 00:13:08.119 "poll_groups": [ 00:13:08.119 { 00:13:08.119 "name": "nvmf_tgt_poll_group_000", 00:13:08.119 "admin_qpairs": 0, 00:13:08.119 "io_qpairs": 0, 00:13:08.119 "current_admin_qpairs": 0, 00:13:08.119 "current_io_qpairs": 0, 00:13:08.119 "pending_bdev_io": 0, 00:13:08.119 "completed_nvme_io": 0, 00:13:08.119 "transports": [ 00:13:08.119 { 00:13:08.119 "trtype": "TCP" 00:13:08.119 } 00:13:08.120 ] 00:13:08.120 }, 00:13:08.120 { 00:13:08.120 "name": "nvmf_tgt_poll_group_001", 00:13:08.120 "admin_qpairs": 0, 00:13:08.120 "io_qpairs": 0, 00:13:08.120 "current_admin_qpairs": 0, 00:13:08.120 "current_io_qpairs": 0, 00:13:08.120 "pending_bdev_io": 0, 00:13:08.120 "completed_nvme_io": 0, 00:13:08.120 "transports": [ 00:13:08.120 { 00:13:08.120 "trtype": "TCP" 00:13:08.120 } 00:13:08.120 ] 00:13:08.120 }, 00:13:08.120 { 00:13:08.120 "name": "nvmf_tgt_poll_group_002", 00:13:08.120 "admin_qpairs": 0, 00:13:08.120 "io_qpairs": 0, 00:13:08.120 "current_admin_qpairs": 0, 00:13:08.120 "current_io_qpairs": 0, 00:13:08.120 "pending_bdev_io": 0, 00:13:08.120 "completed_nvme_io": 0, 00:13:08.120 "transports": [ 00:13:08.120 { 00:13:08.120 "trtype": "TCP" 00:13:08.120 } 00:13:08.120 ] 00:13:08.120 }, 00:13:08.120 { 00:13:08.120 "name": "nvmf_tgt_poll_group_003", 00:13:08.120 "admin_qpairs": 0, 00:13:08.120 "io_qpairs": 0, 00:13:08.120 "current_admin_qpairs": 0, 00:13:08.120 "current_io_qpairs": 0, 00:13:08.120 "pending_bdev_io": 0, 00:13:08.120 "completed_nvme_io": 0, 00:13:08.120 "transports": [ 00:13:08.120 { 00:13:08.120 "trtype": "TCP" 00:13:08.120 } 00:13:08.120 ] 00:13:08.120 } 00:13:08.120 ] 00:13:08.120 }' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # jsum '.poll_groups[].admin_qpairs' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # (( 0 == 0 )) 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # jsum '.poll_groups[].io_qpairs' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:13:08.120 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # (( 0 == 0 )) 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@38 -- # '[' rdma == tcp ']' 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@46 -- # MALLOC_BDEV_SIZE=64 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@47 -- # MALLOC_BLOCK_SIZE=512 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@49 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.380 Malloc1 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@52 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@54 -- # rpc_cmd nvmf_subsystem_allow_any_host -d nqn.2016-06.io.spdk:cnode1 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@55 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.380 [2024-11-26 19:09:20.829124] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@58 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -a 10.0.0.2 -s 4420 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # local es=0 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -a 10.0.0.2 -s 4420 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@640 -- # local arg=nvme 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -t nvme 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # type -P nvme 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # arg=/usr/sbin/nvme 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # [[ -x /usr/sbin/nvme ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -a 10.0.0.2 -s 4420 00:13:08.380 [2024-11-26 19:09:20.865894] ctrlr.c: 825:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396' 00:13:08.380 Failed to write to /dev/nvme-fabrics: Input/output error 00:13:08.380 could not add new controller: failed to write to nvme-fabrics device 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # es=1 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@61 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.380 19:09:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@62 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:10.292 19:09:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@63 -- # waitforserial SPDKISFASTANDAWESOME 00:13:10.292 19:09:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:13:10.292 19:09:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:10.292 19:09:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:10.292 19:09:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@64 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:12.201 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@65 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@68 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@69 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # local es=0 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@640 -- # local arg=nvme 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -t nvme 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # type -P nvme 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # arg=/usr/sbin/nvme 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # [[ -x /usr/sbin/nvme ]] 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:12.201 [2024-11-26 19:09:24.632247] ctrlr.c: 825:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396' 00:13:12.201 Failed to write to /dev/nvme-fabrics: Input/output error 00:13:12.201 could not add new controller: failed to write to nvme-fabrics device 00:13:12.201 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # es=1 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@72 -- # rpc_cmd nvmf_subsystem_allow_any_host -e nqn.2016-06.io.spdk:cnode1 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.202 19:09:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@73 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:13.585 19:09:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@74 -- # waitforserial SPDKISFASTANDAWESOME 00:13:13.585 19:09:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:13:13.585 19:09:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:13.585 19:09:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:13.585 19:09:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@75 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:16.129 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@76 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@78 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # seq 1 5 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:16.129 [2024-11-26 19:09:28.365628] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.129 19:09:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:17.522 19:09:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:13:17.522 19:09:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:13:17.523 19:09:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:17.523 19:09:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:17.523 19:09:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:13:19.436 19:09:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:19.436 19:09:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:19.436 19:09:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:19.436 19:09:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:19.436 19:09:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:19.436 19:09:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:13:19.436 19:09:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:19.436 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.436 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:19.695 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.695 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:19.695 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.695 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:19.695 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:19.696 [2024-11-26 19:09:32.091317] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.696 19:09:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:21.074 19:09:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:13:21.074 19:09:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:13:21.074 19:09:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:21.074 19:09:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:21.074 19:09:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:13:22.984 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:23.244 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:23.244 [2024-11-26 19:09:35.856806] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.244 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:23.504 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.504 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:23.504 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.504 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:23.504 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.504 19:09:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:24.885 19:09:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:13:24.885 19:09:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:13:24.885 19:09:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:24.885 19:09:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:24.885 19:09:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:13:26.858 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:26.859 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:26.859 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:27.118 [2024-11-26 19:09:39.534917] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.118 19:09:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:28.530 19:09:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:13:28.530 19:09:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:13:28.530 19:09:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:28.530 19:09:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:28.530 19:09:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:31.074 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:31.074 [2024-11-26 19:09:43.387438] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.074 19:09:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:32.456 19:09:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:13:32.456 19:09:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:13:32.456 19:09:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:32.456 19:09:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:32.456 19:09:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:13:34.368 19:09:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:34.629 19:09:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:34.629 19:09:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:34.629 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # seq 1 5 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 [2024-11-26 19:09:47.162577] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.629 [2024-11-26 19:09:47.234759] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.629 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 [2024-11-26 19:09:47.302939] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.891 [2024-11-26 19:09:47.375196] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:34.891 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 [2024-11-26 19:09:47.447416] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # rpc_cmd nvmf_get_stats 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.892 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # stats='{ 00:13:35.153 "tick_rate": 2400000000, 00:13:35.153 "poll_groups": [ 00:13:35.153 { 00:13:35.153 "name": "nvmf_tgt_poll_group_000", 00:13:35.153 "admin_qpairs": 0, 00:13:35.153 "io_qpairs": 224, 00:13:35.153 "current_admin_qpairs": 0, 00:13:35.153 "current_io_qpairs": 0, 00:13:35.153 "pending_bdev_io": 0, 00:13:35.153 "completed_nvme_io": 274, 00:13:35.153 "transports": [ 00:13:35.153 { 00:13:35.153 "trtype": "TCP" 00:13:35.153 } 00:13:35.153 ] 00:13:35.153 }, 00:13:35.153 { 00:13:35.153 "name": "nvmf_tgt_poll_group_001", 00:13:35.153 "admin_qpairs": 1, 00:13:35.153 "io_qpairs": 223, 00:13:35.153 "current_admin_qpairs": 0, 00:13:35.153 "current_io_qpairs": 0, 00:13:35.153 "pending_bdev_io": 0, 00:13:35.153 "completed_nvme_io": 517, 00:13:35.153 "transports": [ 00:13:35.153 { 00:13:35.153 "trtype": "TCP" 00:13:35.153 } 00:13:35.153 ] 00:13:35.153 }, 00:13:35.153 { 00:13:35.153 "name": "nvmf_tgt_poll_group_002", 00:13:35.153 "admin_qpairs": 6, 00:13:35.153 "io_qpairs": 218, 00:13:35.153 "current_admin_qpairs": 0, 00:13:35.153 "current_io_qpairs": 0, 00:13:35.153 "pending_bdev_io": 0, 00:13:35.153 "completed_nvme_io": 223, 00:13:35.153 "transports": [ 00:13:35.153 { 00:13:35.153 "trtype": "TCP" 00:13:35.153 } 00:13:35.153 ] 00:13:35.153 }, 00:13:35.153 { 00:13:35.153 "name": "nvmf_tgt_poll_group_003", 00:13:35.153 "admin_qpairs": 0, 00:13:35.153 "io_qpairs": 224, 00:13:35.153 "current_admin_qpairs": 0, 00:13:35.153 "current_io_qpairs": 0, 00:13:35.153 "pending_bdev_io": 0, 00:13:35.153 "completed_nvme_io": 225, 00:13:35.153 "transports": [ 00:13:35.153 { 00:13:35.153 "trtype": "TCP" 00:13:35.153 } 00:13:35.153 ] 00:13:35.153 } 00:13:35.153 ] 00:13:35.153 }' 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # jsum '.poll_groups[].admin_qpairs' 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # (( 7 > 0 )) 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # jsum '.poll_groups[].io_qpairs' 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:13:35.153 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # (( 889 > 0 )) 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@115 -- # '[' rdma == tcp ']' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@123 -- # nvmftestfini 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@516 -- # nvmfcleanup 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@121 -- # sync 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@124 -- # set +e 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:35.154 rmmod nvme_tcp 00:13:35.154 rmmod nvme_fabrics 00:13:35.154 rmmod nvme_keyring 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@128 -- # set -e 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@129 -- # return 0 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@517 -- # '[' -n 3766725 ']' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@518 -- # killprocess 3766725 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@954 -- # '[' -z 3766725 ']' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@958 -- # kill -0 3766725 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@959 -- # uname 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3766725 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3766725' 00:13:35.154 killing process with pid 3766725 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@973 -- # kill 3766725 00:13:35.154 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@978 -- # wait 3766725 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@297 -- # iptr 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@791 -- # iptables-restore 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@791 -- # iptables-save 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:35.414 19:09:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:37.958 19:09:49 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:37.958 00:13:37.958 real 0m38.997s 00:13:37.958 user 1m54.575s 00:13:37.958 sys 0m8.505s 00:13:37.958 19:09:49 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:37.958 19:09:49 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:13:37.958 ************************************ 00:13:37.958 END TEST nvmf_rpc 00:13:37.958 ************************************ 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@23 -- # run_test nvmf_invalid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:13:37.958 ************************************ 00:13:37.958 START TEST nvmf_invalid 00:13:37.958 ************************************ 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:13:37.958 * Looking for test storage... 00:13:37.958 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1693 -- # lcov --version 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # IFS=.-: 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # read -ra ver1 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # IFS=.-: 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # read -ra ver2 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@338 -- # local 'op=<' 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@340 -- # ver1_l=2 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@341 -- # ver2_l=1 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@344 -- # case "$op" in 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@345 -- # : 1 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # decimal 1 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=1 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 1 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # ver1[v]=1 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # decimal 2 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=2 00:13:37.958 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 2 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # ver2[v]=2 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # return 0 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:13:37.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:37.959 --rc genhtml_branch_coverage=1 00:13:37.959 --rc genhtml_function_coverage=1 00:13:37.959 --rc genhtml_legend=1 00:13:37.959 --rc geninfo_all_blocks=1 00:13:37.959 --rc geninfo_unexecuted_blocks=1 00:13:37.959 00:13:37.959 ' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:13:37.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:37.959 --rc genhtml_branch_coverage=1 00:13:37.959 --rc genhtml_function_coverage=1 00:13:37.959 --rc genhtml_legend=1 00:13:37.959 --rc geninfo_all_blocks=1 00:13:37.959 --rc geninfo_unexecuted_blocks=1 00:13:37.959 00:13:37.959 ' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:13:37.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:37.959 --rc genhtml_branch_coverage=1 00:13:37.959 --rc genhtml_function_coverage=1 00:13:37.959 --rc genhtml_legend=1 00:13:37.959 --rc geninfo_all_blocks=1 00:13:37.959 --rc geninfo_unexecuted_blocks=1 00:13:37.959 00:13:37.959 ' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:13:37.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:37.959 --rc genhtml_branch_coverage=1 00:13:37.959 --rc genhtml_function_coverage=1 00:13:37.959 --rc genhtml_legend=1 00:13:37.959 --rc geninfo_all_blocks=1 00:13:37.959 --rc geninfo_unexecuted_blocks=1 00:13:37.959 00:13:37.959 ' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # uname -s 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@15 -- # shopt -s extglob 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@5 -- # export PATH 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@51 -- # : 0 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:37.959 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@11 -- # multi_target_rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@12 -- # rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@14 -- # target=foobar 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@16 -- # RANDOM=0 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@34 -- # nvmftestinit 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@476 -- # prepare_net_devs 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@438 -- # local -g is_hw=no 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # remove_spdk_ns 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@309 -- # xtrace_disable 00:13:37.959 19:09:50 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # pci_devs=() 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # net_devs=() 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # e810=() 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # local -ga e810 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # x722=() 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # local -ga x722 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # mlx=() 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # local -ga mlx 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:13:46.119 Found 0000:31:00.0 (0x8086 - 0x159b) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:13:46.119 Found 0000:31:00.1 (0x8086 - 0x159b) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:13:46.119 Found net devices under 0000:31:00.0: cvl_0_0 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:13:46.119 Found net devices under 0000:31:00.1: cvl_0_1 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # is_hw=yes 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:46.119 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:46.119 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.642 ms 00:13:46.119 00:13:46.119 --- 10.0.0.2 ping statistics --- 00:13:46.119 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:46.119 rtt min/avg/max/mdev = 0.642/0.642/0.642/0.000 ms 00:13:46.119 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:46.119 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:46.119 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.188 ms 00:13:46.119 00:13:46.119 --- 10.0.0.1 ping statistics --- 00:13:46.120 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:46.120 rtt min/avg/max/mdev = 0.188/0.188/0.188/0.000 ms 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@450 -- # return 0 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@35 -- # nvmfappstart -m 0xF 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@726 -- # xtrace_disable 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@509 -- # nvmfpid=3776989 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@510 -- # waitforlisten 3776989 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@835 -- # '[' -z 3776989 ']' 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:46.120 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:46.120 19:09:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:13:46.120 [2024-11-26 19:09:58.511383] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:13:46.120 [2024-11-26 19:09:58.511446] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:46.120 [2024-11-26 19:09:58.605719] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:46.120 [2024-11-26 19:09:58.647084] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:46.120 [2024-11-26 19:09:58.647125] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:46.120 [2024-11-26 19:09:58.647133] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:46.120 [2024-11-26 19:09:58.647141] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:46.120 [2024-11-26 19:09:58.647146] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:46.120 [2024-11-26 19:09:58.648797] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:46.120 [2024-11-26 19:09:58.648931] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:46.120 [2024-11-26 19:09:58.649222] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:46.120 [2024-11-26 19:09:58.649224] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@868 -- # return 0 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@732 -- # xtrace_disable 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@37 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -t foobar nqn.2016-06.io.spdk:cnode27246 00:13:47.061 [2024-11-26 19:09:59.526254] nvmf_rpc.c: 396:rpc_nvmf_create_subsystem: *ERROR*: Unable to find target foobar 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # out='request: 00:13:47.061 { 00:13:47.061 "nqn": "nqn.2016-06.io.spdk:cnode27246", 00:13:47.061 "tgt_name": "foobar", 00:13:47.061 "method": "nvmf_create_subsystem", 00:13:47.061 "req_id": 1 00:13:47.061 } 00:13:47.061 Got JSON-RPC error response 00:13:47.061 response: 00:13:47.061 { 00:13:47.061 "code": -32603, 00:13:47.061 "message": "Unable to find target foobar" 00:13:47.061 }' 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@41 -- # [[ request: 00:13:47.061 { 00:13:47.061 "nqn": "nqn.2016-06.io.spdk:cnode27246", 00:13:47.061 "tgt_name": "foobar", 00:13:47.061 "method": "nvmf_create_subsystem", 00:13:47.061 "req_id": 1 00:13:47.061 } 00:13:47.061 Got JSON-RPC error response 00:13:47.061 response: 00:13:47.061 { 00:13:47.061 "code": -32603, 00:13:47.061 "message": "Unable to find target foobar" 00:13:47.061 } == *\U\n\a\b\l\e\ \t\o\ \f\i\n\d\ \t\a\r\g\e\t* ]] 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # echo -e '\x1f' 00:13:47.061 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s $'SPDKISFASTANDAWESOME\037' nqn.2016-06.io.spdk:cnode25422 00:13:47.322 [2024-11-26 19:09:59.718951] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode25422: invalid serial number 'SPDKISFASTANDAWESOME' 00:13:47.322 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # out='request: 00:13:47.322 { 00:13:47.322 "nqn": "nqn.2016-06.io.spdk:cnode25422", 00:13:47.322 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:13:47.322 "method": "nvmf_create_subsystem", 00:13:47.322 "req_id": 1 00:13:47.322 } 00:13:47.322 Got JSON-RPC error response 00:13:47.322 response: 00:13:47.322 { 00:13:47.322 "code": -32602, 00:13:47.322 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:13:47.322 }' 00:13:47.322 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@46 -- # [[ request: 00:13:47.322 { 00:13:47.322 "nqn": "nqn.2016-06.io.spdk:cnode25422", 00:13:47.322 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:13:47.322 "method": "nvmf_create_subsystem", 00:13:47.322 "req_id": 1 00:13:47.322 } 00:13:47.322 Got JSON-RPC error response 00:13:47.322 response: 00:13:47.322 { 00:13:47.322 "code": -32602, 00:13:47.322 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:13:47.322 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:13:47.322 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # echo -e '\x1f' 00:13:47.322 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d $'SPDK_Controller\037' nqn.2016-06.io.spdk:cnode1668 00:13:47.322 [2024-11-26 19:09:59.911513] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode1668: invalid model number 'SPDK_Controller' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # out='request: 00:13:47.583 { 00:13:47.583 "nqn": "nqn.2016-06.io.spdk:cnode1668", 00:13:47.583 "model_number": "SPDK_Controller\u001f", 00:13:47.583 "method": "nvmf_create_subsystem", 00:13:47.583 "req_id": 1 00:13:47.583 } 00:13:47.583 Got JSON-RPC error response 00:13:47.583 response: 00:13:47.583 { 00:13:47.583 "code": -32602, 00:13:47.583 "message": "Invalid MN SPDK_Controller\u001f" 00:13:47.583 }' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@51 -- # [[ request: 00:13:47.583 { 00:13:47.583 "nqn": "nqn.2016-06.io.spdk:cnode1668", 00:13:47.583 "model_number": "SPDK_Controller\u001f", 00:13:47.583 "method": "nvmf_create_subsystem", 00:13:47.583 "req_id": 1 00:13:47.583 } 00:13:47.583 Got JSON-RPC error response 00:13:47.583 response: 00:13:47.583 { 00:13:47.583 "code": -32602, 00:13:47.583 "message": "Invalid MN SPDK_Controller\u001f" 00:13:47.583 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # gen_random_s 21 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@19 -- # local length=21 ll 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 80 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x50' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=P 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 119 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x77' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=w 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 85 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x55' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=U 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 50 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x32' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=2 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 42 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2a' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='*' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 116 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x74' 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=t 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.583 19:09:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 94 00:13:47.583 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5e' 00:13:47.583 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='^' 00:13:47.583 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.583 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 98 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x62' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=b 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 66 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x42' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=B 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 44 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2c' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=, 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 85 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x55' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=U 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 68 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x44' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=D 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 35 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x23' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='#' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 72 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x48' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=H 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 39 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x27' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=\' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 41 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x29' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=')' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 39 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x27' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=\' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 81 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x51' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Q 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 83 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x53' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=S 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 72 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x48' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=H 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 83 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x53' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=S 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@28 -- # [[ P == \- ]] 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@31 -- # echo 'PwU2*t^bB,UD#H'\'')'\''QSHS' 00:13:47.584 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s 'PwU2*t^bB,UD#H'\'')'\''QSHS' nqn.2016-06.io.spdk:cnode17560 00:13:47.845 [2024-11-26 19:10:00.276723] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode17560: invalid serial number 'PwU2*t^bB,UD#H')'QSHS' 00:13:47.845 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # out='request: 00:13:47.845 { 00:13:47.845 "nqn": "nqn.2016-06.io.spdk:cnode17560", 00:13:47.845 "serial_number": "PwU2*t^bB,UD#H'\'')'\''QSHS", 00:13:47.845 "method": "nvmf_create_subsystem", 00:13:47.845 "req_id": 1 00:13:47.845 } 00:13:47.845 Got JSON-RPC error response 00:13:47.845 response: 00:13:47.845 { 00:13:47.845 "code": -32602, 00:13:47.845 "message": "Invalid SN PwU2*t^bB,UD#H'\'')'\''QSHS" 00:13:47.845 }' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@55 -- # [[ request: 00:13:47.846 { 00:13:47.846 "nqn": "nqn.2016-06.io.spdk:cnode17560", 00:13:47.846 "serial_number": "PwU2*t^bB,UD#H')'QSHS", 00:13:47.846 "method": "nvmf_create_subsystem", 00:13:47.846 "req_id": 1 00:13:47.846 } 00:13:47.846 Got JSON-RPC error response 00:13:47.846 response: 00:13:47.846 { 00:13:47.846 "code": -32602, 00:13:47.846 "message": "Invalid SN PwU2*t^bB,UD#H')'QSHS" 00:13:47.846 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # gen_random_s 41 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@19 -- # local length=41 ll 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 51 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x33' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=3 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 114 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x72' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=r 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 66 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x42' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=B 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 55 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x37' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=7 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 55 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x37' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=7 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 122 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7a' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=z 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 69 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x45' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=E 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 115 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x73' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=s 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 96 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x60' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='`' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 105 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x69' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=i 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 46 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2e' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=. 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 115 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x73' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=s 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 71 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x47' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=G 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 42 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2a' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='*' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 107 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6b' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=k 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 56 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x38' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=8 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 67 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x43' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=C 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 115 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x73' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=s 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 43 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2b' 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=+ 00:13:47.846 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:47.847 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 104 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x68' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=h 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 100 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x64' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=d 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 43 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2b' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=+ 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 95 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5f' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=_ 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 77 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4d' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=M 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 63 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3f' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='?' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 45 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2d' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=- 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 52 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x34' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=4 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 53 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x35' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=5 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 97 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x61' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=a 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 100 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x64' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=d 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 58 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3a' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=: 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 47 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2f' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=/ 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 96 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x60' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='`' 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 63 00:13:48.108 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3f' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='?' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 45 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2d' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=- 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 112 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x70' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=p 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 117 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x75' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=u 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 81 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x51' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Q 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 121 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x79' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=y 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 81 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x51' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Q 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 101 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x65' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=e 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@28 -- # [[ 3 == \- ]] 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@31 -- # echo '3rB77zEs`i.sG*k8Cs+hd+_M?-45ad:/`?-puQyQe' 00:13:48.109 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d '3rB77zEs`i.sG*k8Cs+hd+_M?-45ad:/`?-puQyQe' nqn.2016-06.io.spdk:cnode3819 00:13:48.369 [2024-11-26 19:10:00.798462] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode3819: invalid model number '3rB77zEs`i.sG*k8Cs+hd+_M?-45ad:/`?-puQyQe' 00:13:48.369 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # out='request: 00:13:48.369 { 00:13:48.369 "nqn": "nqn.2016-06.io.spdk:cnode3819", 00:13:48.369 "model_number": "3rB77zEs`i.sG*k8Cs+hd+_M?-45ad:/`?-puQyQe", 00:13:48.369 "method": "nvmf_create_subsystem", 00:13:48.369 "req_id": 1 00:13:48.369 } 00:13:48.369 Got JSON-RPC error response 00:13:48.369 response: 00:13:48.369 { 00:13:48.369 "code": -32602, 00:13:48.369 "message": "Invalid MN 3rB77zEs`i.sG*k8Cs+hd+_M?-45ad:/`?-puQyQe" 00:13:48.369 }' 00:13:48.369 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@59 -- # [[ request: 00:13:48.369 { 00:13:48.369 "nqn": "nqn.2016-06.io.spdk:cnode3819", 00:13:48.369 "model_number": "3rB77zEs`i.sG*k8Cs+hd+_M?-45ad:/`?-puQyQe", 00:13:48.369 "method": "nvmf_create_subsystem", 00:13:48.369 "req_id": 1 00:13:48.369 } 00:13:48.369 Got JSON-RPC error response 00:13:48.369 response: 00:13:48.369 { 00:13:48.369 "code": -32602, 00:13:48.369 "message": "Invalid MN 3rB77zEs`i.sG*k8Cs+hd+_M?-45ad:/`?-puQyQe" 00:13:48.369 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:13:48.369 19:10:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport --trtype tcp 00:13:48.369 [2024-11-26 19:10:00.983151] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:48.631 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode -s SPDK001 -a 00:13:48.631 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@64 -- # [[ tcp == \T\C\P ]] 00:13:48.631 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # echo '' 00:13:48.631 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # head -n 1 00:13:48.631 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # IP= 00:13:48.631 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode -t tcp -a '' -s 4421 00:13:48.891 [2024-11-26 19:10:01.368279] nvmf_rpc.c: 783:nvmf_rpc_listen_paused: *ERROR*: Unable to remove listener, rc -2 00:13:48.891 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # out='request: 00:13:48.891 { 00:13:48.891 "nqn": "nqn.2016-06.io.spdk:cnode", 00:13:48.891 "listen_address": { 00:13:48.891 "trtype": "tcp", 00:13:48.891 "traddr": "", 00:13:48.891 "trsvcid": "4421" 00:13:48.891 }, 00:13:48.891 "method": "nvmf_subsystem_remove_listener", 00:13:48.891 "req_id": 1 00:13:48.891 } 00:13:48.891 Got JSON-RPC error response 00:13:48.891 response: 00:13:48.891 { 00:13:48.891 "code": -32602, 00:13:48.891 "message": "Invalid parameters" 00:13:48.891 }' 00:13:48.891 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@70 -- # [[ request: 00:13:48.891 { 00:13:48.891 "nqn": "nqn.2016-06.io.spdk:cnode", 00:13:48.891 "listen_address": { 00:13:48.891 "trtype": "tcp", 00:13:48.891 "traddr": "", 00:13:48.891 "trsvcid": "4421" 00:13:48.891 }, 00:13:48.891 "method": "nvmf_subsystem_remove_listener", 00:13:48.891 "req_id": 1 00:13:48.891 } 00:13:48.891 Got JSON-RPC error response 00:13:48.891 response: 00:13:48.891 { 00:13:48.892 "code": -32602, 00:13:48.892 "message": "Invalid parameters" 00:13:48.892 } != *\U\n\a\b\l\e\ \t\o\ \s\t\o\p\ \l\i\s\t\e\n\e\r\.* ]] 00:13:48.892 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode18117 -i 0 00:13:49.152 [2024-11-26 19:10:01.552823] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode18117: invalid cntlid range [0-65519] 00:13:49.152 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # out='request: 00:13:49.152 { 00:13:49.152 "nqn": "nqn.2016-06.io.spdk:cnode18117", 00:13:49.152 "min_cntlid": 0, 00:13:49.152 "method": "nvmf_create_subsystem", 00:13:49.152 "req_id": 1 00:13:49.152 } 00:13:49.152 Got JSON-RPC error response 00:13:49.152 response: 00:13:49.152 { 00:13:49.152 "code": -32602, 00:13:49.152 "message": "Invalid cntlid range [0-65519]" 00:13:49.152 }' 00:13:49.152 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@74 -- # [[ request: 00:13:49.152 { 00:13:49.152 "nqn": "nqn.2016-06.io.spdk:cnode18117", 00:13:49.152 "min_cntlid": 0, 00:13:49.152 "method": "nvmf_create_subsystem", 00:13:49.152 "req_id": 1 00:13:49.152 } 00:13:49.152 Got JSON-RPC error response 00:13:49.152 response: 00:13:49.152 { 00:13:49.152 "code": -32602, 00:13:49.152 "message": "Invalid cntlid range [0-65519]" 00:13:49.152 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:13:49.152 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode13638 -i 65520 00:13:49.152 [2024-11-26 19:10:01.737400] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode13638: invalid cntlid range [65520-65519] 00:13:49.152 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # out='request: 00:13:49.152 { 00:13:49.152 "nqn": "nqn.2016-06.io.spdk:cnode13638", 00:13:49.152 "min_cntlid": 65520, 00:13:49.152 "method": "nvmf_create_subsystem", 00:13:49.152 "req_id": 1 00:13:49.152 } 00:13:49.152 Got JSON-RPC error response 00:13:49.152 response: 00:13:49.152 { 00:13:49.152 "code": -32602, 00:13:49.152 "message": "Invalid cntlid range [65520-65519]" 00:13:49.152 }' 00:13:49.152 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@76 -- # [[ request: 00:13:49.152 { 00:13:49.152 "nqn": "nqn.2016-06.io.spdk:cnode13638", 00:13:49.152 "min_cntlid": 65520, 00:13:49.152 "method": "nvmf_create_subsystem", 00:13:49.152 "req_id": 1 00:13:49.152 } 00:13:49.152 Got JSON-RPC error response 00:13:49.152 response: 00:13:49.152 { 00:13:49.152 "code": -32602, 00:13:49.152 "message": "Invalid cntlid range [65520-65519]" 00:13:49.152 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:13:49.152 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode12565 -I 0 00:13:49.412 [2024-11-26 19:10:01.925987] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode12565: invalid cntlid range [1-0] 00:13:49.412 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # out='request: 00:13:49.412 { 00:13:49.412 "nqn": "nqn.2016-06.io.spdk:cnode12565", 00:13:49.412 "max_cntlid": 0, 00:13:49.412 "method": "nvmf_create_subsystem", 00:13:49.412 "req_id": 1 00:13:49.412 } 00:13:49.412 Got JSON-RPC error response 00:13:49.412 response: 00:13:49.412 { 00:13:49.412 "code": -32602, 00:13:49.412 "message": "Invalid cntlid range [1-0]" 00:13:49.412 }' 00:13:49.412 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@78 -- # [[ request: 00:13:49.412 { 00:13:49.412 "nqn": "nqn.2016-06.io.spdk:cnode12565", 00:13:49.412 "max_cntlid": 0, 00:13:49.412 "method": "nvmf_create_subsystem", 00:13:49.412 "req_id": 1 00:13:49.412 } 00:13:49.412 Got JSON-RPC error response 00:13:49.412 response: 00:13:49.412 { 00:13:49.412 "code": -32602, 00:13:49.412 "message": "Invalid cntlid range [1-0]" 00:13:49.412 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:13:49.412 19:10:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode32411 -I 65520 00:13:49.673 [2024-11-26 19:10:02.114577] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode32411: invalid cntlid range [1-65520] 00:13:49.673 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # out='request: 00:13:49.673 { 00:13:49.673 "nqn": "nqn.2016-06.io.spdk:cnode32411", 00:13:49.673 "max_cntlid": 65520, 00:13:49.673 "method": "nvmf_create_subsystem", 00:13:49.673 "req_id": 1 00:13:49.673 } 00:13:49.673 Got JSON-RPC error response 00:13:49.673 response: 00:13:49.673 { 00:13:49.673 "code": -32602, 00:13:49.673 "message": "Invalid cntlid range [1-65520]" 00:13:49.673 }' 00:13:49.673 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@80 -- # [[ request: 00:13:49.673 { 00:13:49.673 "nqn": "nqn.2016-06.io.spdk:cnode32411", 00:13:49.673 "max_cntlid": 65520, 00:13:49.673 "method": "nvmf_create_subsystem", 00:13:49.673 "req_id": 1 00:13:49.673 } 00:13:49.673 Got JSON-RPC error response 00:13:49.673 response: 00:13:49.673 { 00:13:49.673 "code": -32602, 00:13:49.673 "message": "Invalid cntlid range [1-65520]" 00:13:49.673 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:13:49.673 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode7471 -i 6 -I 5 00:13:49.934 [2024-11-26 19:10:02.303164] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode7471: invalid cntlid range [6-5] 00:13:49.934 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # out='request: 00:13:49.934 { 00:13:49.934 "nqn": "nqn.2016-06.io.spdk:cnode7471", 00:13:49.934 "min_cntlid": 6, 00:13:49.934 "max_cntlid": 5, 00:13:49.934 "method": "nvmf_create_subsystem", 00:13:49.934 "req_id": 1 00:13:49.934 } 00:13:49.934 Got JSON-RPC error response 00:13:49.934 response: 00:13:49.934 { 00:13:49.934 "code": -32602, 00:13:49.934 "message": "Invalid cntlid range [6-5]" 00:13:49.934 }' 00:13:49.934 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@84 -- # [[ request: 00:13:49.934 { 00:13:49.934 "nqn": "nqn.2016-06.io.spdk:cnode7471", 00:13:49.934 "min_cntlid": 6, 00:13:49.934 "max_cntlid": 5, 00:13:49.934 "method": "nvmf_create_subsystem", 00:13:49.934 "req_id": 1 00:13:49.934 } 00:13:49.934 Got JSON-RPC error response 00:13:49.934 response: 00:13:49.934 { 00:13:49.934 "code": -32602, 00:13:49.934 "message": "Invalid cntlid range [6-5]" 00:13:49.934 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:13:49.934 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target --name foobar 00:13:49.934 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # out='request: 00:13:49.934 { 00:13:49.934 "name": "foobar", 00:13:49.934 "method": "nvmf_delete_target", 00:13:49.934 "req_id": 1 00:13:49.934 } 00:13:49.934 Got JSON-RPC error response 00:13:49.934 response: 00:13:49.934 { 00:13:49.934 "code": -32602, 00:13:49.934 "message": "The specified target doesn'\''t exist, cannot delete it." 00:13:49.934 }' 00:13:49.934 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@88 -- # [[ request: 00:13:49.934 { 00:13:49.934 "name": "foobar", 00:13:49.934 "method": "nvmf_delete_target", 00:13:49.934 "req_id": 1 00:13:49.934 } 00:13:49.934 Got JSON-RPC error response 00:13:49.935 response: 00:13:49.935 { 00:13:49.935 "code": -32602, 00:13:49.935 "message": "The specified target doesn't exist, cannot delete it." 00:13:49.935 } == *\T\h\e\ \s\p\e\c\i\f\i\e\d\ \t\a\r\g\e\t\ \d\o\e\s\n\'\t\ \e\x\i\s\t\,\ \c\a\n\n\o\t\ \d\e\l\e\t\e\ \i\t\.* ]] 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@90 -- # trap - SIGINT SIGTERM EXIT 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@91 -- # nvmftestfini 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@516 -- # nvmfcleanup 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@121 -- # sync 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@124 -- # set +e 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:49.935 rmmod nvme_tcp 00:13:49.935 rmmod nvme_fabrics 00:13:49.935 rmmod nvme_keyring 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@128 -- # set -e 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@129 -- # return 0 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@517 -- # '[' -n 3776989 ']' 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@518 -- # killprocess 3776989 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@954 -- # '[' -z 3776989 ']' 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@958 -- # kill -0 3776989 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@959 -- # uname 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:49.935 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3776989 00:13:50.196 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:50.196 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:50.196 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3776989' 00:13:50.196 killing process with pid 3776989 00:13:50.196 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@973 -- # kill 3776989 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@978 -- # wait 3776989 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@297 -- # iptr 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@791 -- # iptables-restore 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@791 -- # iptables-save 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:50.197 19:10:02 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:52.743 00:13:52.743 real 0m14.719s 00:13:52.743 user 0m20.922s 00:13:52.743 sys 0m7.099s 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:13:52.743 ************************************ 00:13:52.743 END TEST nvmf_invalid 00:13:52.743 ************************************ 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@24 -- # run_test nvmf_connect_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:13:52.743 ************************************ 00:13:52.743 START TEST nvmf_connect_stress 00:13:52.743 ************************************ 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:13:52.743 * Looking for test storage... 00:13:52.743 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1693 -- # lcov --version 00:13:52.743 19:10:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # IFS=.-: 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # read -ra ver1 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # IFS=.-: 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # read -ra ver2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@338 -- # local 'op=<' 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@340 -- # ver1_l=2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@341 -- # ver2_l=1 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@344 -- # case "$op" in 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@345 -- # : 1 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # decimal 1 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=1 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 1 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # decimal 2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # return 0 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:13:52.743 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:52.743 --rc genhtml_branch_coverage=1 00:13:52.743 --rc genhtml_function_coverage=1 00:13:52.743 --rc genhtml_legend=1 00:13:52.743 --rc geninfo_all_blocks=1 00:13:52.743 --rc geninfo_unexecuted_blocks=1 00:13:52.743 00:13:52.743 ' 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:13:52.743 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:52.743 --rc genhtml_branch_coverage=1 00:13:52.743 --rc genhtml_function_coverage=1 00:13:52.743 --rc genhtml_legend=1 00:13:52.743 --rc geninfo_all_blocks=1 00:13:52.743 --rc geninfo_unexecuted_blocks=1 00:13:52.743 00:13:52.743 ' 00:13:52.743 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:13:52.744 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:52.744 --rc genhtml_branch_coverage=1 00:13:52.744 --rc genhtml_function_coverage=1 00:13:52.744 --rc genhtml_legend=1 00:13:52.744 --rc geninfo_all_blocks=1 00:13:52.744 --rc geninfo_unexecuted_blocks=1 00:13:52.744 00:13:52.744 ' 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:13:52.744 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:52.744 --rc genhtml_branch_coverage=1 00:13:52.744 --rc genhtml_function_coverage=1 00:13:52.744 --rc genhtml_legend=1 00:13:52.744 --rc geninfo_all_blocks=1 00:13:52.744 --rc geninfo_unexecuted_blocks=1 00:13:52.744 00:13:52.744 ' 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # uname -s 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@5 -- # export PATH 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@51 -- # : 0 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:52.744 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@12 -- # nvmftestinit 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@476 -- # prepare_net_devs 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@438 -- # local -g is_hw=no 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # remove_spdk_ns 00:13:52.744 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:52.745 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:52.745 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:52.745 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:13:52.745 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:13:52.745 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:13:52.745 19:10:05 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # net_devs=() 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # e810=() 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # local -ga e810 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # x722=() 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # local -ga x722 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # mlx=() 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:14:00.880 Found 0000:31:00.0 (0x8086 - 0x159b) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:14:00.880 Found 0000:31:00.1 (0x8086 - 0x159b) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:14:00.880 Found net devices under 0000:31:00.0: cvl_0_0 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:14:00.880 Found net devices under 0000:31:00.1: cvl_0_1 00:14:00.880 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # is_hw=yes 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:00.881 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:01.142 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:01.142 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.681 ms 00:14:01.142 00:14:01.142 --- 10.0.0.2 ping statistics --- 00:14:01.142 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:01.142 rtt min/avg/max/mdev = 0.681/0.681/0.681/0.000 ms 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:01.142 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:01.142 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.186 ms 00:14:01.142 00:14:01.142 --- 10.0.0.1 ping statistics --- 00:14:01.142 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:01.142 rtt min/avg/max/mdev = 0.186/0.186/0.186/0.000 ms 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@450 -- # return 0 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@13 -- # nvmfappstart -m 0xE 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@509 -- # nvmfpid=3782792 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@510 -- # waitforlisten 3782792 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@835 -- # '[' -z 3782792 ']' 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:01.142 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:01.142 19:10:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:01.142 [2024-11-26 19:10:13.741085] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:14:01.142 [2024-11-26 19:10:13.741154] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:01.403 [2024-11-26 19:10:13.849271] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:14:01.403 [2024-11-26 19:10:13.900503] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:01.403 [2024-11-26 19:10:13.900556] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:01.403 [2024-11-26 19:10:13.900565] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:01.403 [2024-11-26 19:10:13.900572] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:01.403 [2024-11-26 19:10:13.900579] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:01.403 [2024-11-26 19:10:13.902514] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:01.403 [2024-11-26 19:10:13.902676] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:01.403 [2024-11-26 19:10:13.902676] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:01.973 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:01.973 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@868 -- # return 0 00:14:01.973 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:14:01.973 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:01.973 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:02.234 [2024-11-26 19:10:14.615415] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:02.234 [2024-11-26 19:10:14.639799] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:02.234 NULL1 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@21 -- # PERF_PID=3783028 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@23 -- # rpcs=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/connect_stress/connect_stress -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -t 10 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@25 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # seq 1 20 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.234 19:10:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:02.495 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.495 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:02.495 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:02.495 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.495 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:03.065 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.065 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:03.065 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:03.065 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.065 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:03.325 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.325 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:03.325 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:03.325 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.325 19:10:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:03.585 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.585 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:03.585 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:03.586 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.586 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:03.846 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.846 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:03.846 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:03.846 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.847 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:04.107 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.107 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:04.107 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:04.107 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.107 19:10:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:04.676 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.676 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:04.676 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:04.676 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.676 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:04.937 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.937 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:04.937 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:04.937 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.937 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:05.197 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.197 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:05.197 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:05.197 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.197 19:10:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:05.457 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.457 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:05.457 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:05.457 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.457 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:05.717 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.717 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:05.717 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:05.717 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.718 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:06.288 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.288 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:06.288 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:06.288 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.288 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:06.548 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.548 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:06.548 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:06.548 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.548 19:10:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:06.807 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.807 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:06.807 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:06.807 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.807 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:07.067 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.067 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:07.067 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:07.067 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.067 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:07.692 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.692 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:07.692 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:07.692 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.692 19:10:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.021 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:08.590 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.590 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:08.590 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:08.590 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.590 19:10:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:08.850 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.850 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:08.850 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:08.850 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.850 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:09.111 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.111 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:09.111 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:09.111 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.111 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:09.371 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.371 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:09.371 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:09.371 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.371 19:10:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:09.941 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.941 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:09.941 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:09.941 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.941 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:10.202 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.202 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:10.202 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:10.202 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.202 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:10.464 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.464 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:10.464 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:10.464 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.464 19:10:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:10.725 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.725 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:10.725 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:10.725 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.725 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:10.985 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.985 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:10.985 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:10.985 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.985 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:11.554 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.554 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:11.554 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:11.554 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.554 19:10:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:11.814 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.814 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:11.814 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:11.814 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.814 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:12.075 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.075 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:12.075 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:14:12.075 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.075 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:12.337 Testing NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 3783028 00:14:12.337 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh: line 34: kill: (3783028) - No such process 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@38 -- # wait 3783028 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@39 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@43 -- # nvmftestfini 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@516 -- # nvmfcleanup 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@121 -- # sync 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@124 -- # set +e 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:12.337 rmmod nvme_tcp 00:14:12.337 rmmod nvme_fabrics 00:14:12.337 rmmod nvme_keyring 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@128 -- # set -e 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@129 -- # return 0 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@517 -- # '[' -n 3782792 ']' 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@518 -- # killprocess 3782792 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@954 -- # '[' -z 3782792 ']' 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@958 -- # kill -0 3782792 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@959 -- # uname 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:12.337 19:10:24 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3782792 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3782792' 00:14:12.598 killing process with pid 3782792 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@973 -- # kill 3782792 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@978 -- # wait 3782792 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@297 -- # iptr 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@791 -- # iptables-restore 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@791 -- # iptables-save 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:12.598 19:10:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:14:15.143 00:14:15.143 real 0m22.345s 00:14:15.143 user 0m42.609s 00:14:15.143 sys 0m10.031s 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:14:15.143 ************************************ 00:14:15.143 END TEST nvmf_connect_stress 00:14:15.143 ************************************ 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@25 -- # run_test nvmf_fused_ordering /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:14:15.143 ************************************ 00:14:15.143 START TEST nvmf_fused_ordering 00:14:15.143 ************************************ 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:14:15.143 * Looking for test storage... 00:14:15.143 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1693 -- # lcov --version 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # IFS=.-: 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # read -ra ver1 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # IFS=.-: 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # read -ra ver2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@338 -- # local 'op=<' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@340 -- # ver1_l=2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@341 -- # ver2_l=1 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@344 -- # case "$op" in 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@345 -- # : 1 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # decimal 1 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=1 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 1 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # ver1[v]=1 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # decimal 2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # ver2[v]=2 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # return 0 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:14:15.143 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:15.143 --rc genhtml_branch_coverage=1 00:14:15.143 --rc genhtml_function_coverage=1 00:14:15.143 --rc genhtml_legend=1 00:14:15.143 --rc geninfo_all_blocks=1 00:14:15.143 --rc geninfo_unexecuted_blocks=1 00:14:15.143 00:14:15.143 ' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:14:15.143 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:15.143 --rc genhtml_branch_coverage=1 00:14:15.143 --rc genhtml_function_coverage=1 00:14:15.143 --rc genhtml_legend=1 00:14:15.143 --rc geninfo_all_blocks=1 00:14:15.143 --rc geninfo_unexecuted_blocks=1 00:14:15.143 00:14:15.143 ' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:14:15.143 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:15.143 --rc genhtml_branch_coverage=1 00:14:15.143 --rc genhtml_function_coverage=1 00:14:15.143 --rc genhtml_legend=1 00:14:15.143 --rc geninfo_all_blocks=1 00:14:15.143 --rc geninfo_unexecuted_blocks=1 00:14:15.143 00:14:15.143 ' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:14:15.143 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:15.143 --rc genhtml_branch_coverage=1 00:14:15.143 --rc genhtml_function_coverage=1 00:14:15.143 --rc genhtml_legend=1 00:14:15.143 --rc geninfo_all_blocks=1 00:14:15.143 --rc geninfo_unexecuted_blocks=1 00:14:15.143 00:14:15.143 ' 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # uname -s 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:15.143 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@15 -- # shopt -s extglob 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@5 -- # export PATH 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@51 -- # : 0 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:15.144 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@12 -- # nvmftestinit 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@476 -- # prepare_net_devs 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@438 -- # local -g is_hw=no 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # remove_spdk_ns 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@309 -- # xtrace_disable 00:14:15.144 19:10:27 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # pci_devs=() 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # net_devs=() 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # e810=() 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # local -ga e810 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # x722=() 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # local -ga x722 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # mlx=() 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # local -ga mlx 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:14:23.284 Found 0000:31:00.0 (0x8086 - 0x159b) 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:23.284 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:14:23.285 Found 0000:31:00.1 (0x8086 - 0x159b) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:14:23.285 Found net devices under 0000:31:00.0: cvl_0_0 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:14:23.285 Found net devices under 0000:31:00.1: cvl_0_1 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # is_hw=yes 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:23.285 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:23.548 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:23.548 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.571 ms 00:14:23.548 00:14:23.548 --- 10.0.0.2 ping statistics --- 00:14:23.548 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:23.548 rtt min/avg/max/mdev = 0.571/0.571/0.571/0.000 ms 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:23.548 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:23.548 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.307 ms 00:14:23.548 00:14:23.548 --- 10.0.0.1 ping statistics --- 00:14:23.548 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:23.548 rtt min/avg/max/mdev = 0.307/0.307/0.307/0.000 ms 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@450 -- # return 0 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:14:23.548 19:10:35 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@13 -- # nvmfappstart -m 0x2 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@509 -- # nvmfpid=3789862 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@510 -- # waitforlisten 3789862 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@835 -- # '[' -z 3789862 ']' 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:23.548 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:23.548 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:23.548 [2024-11-26 19:10:36.087136] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:14:23.548 [2024-11-26 19:10:36.087202] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:23.809 [2024-11-26 19:10:36.193677] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:23.809 [2024-11-26 19:10:36.243011] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:23.809 [2024-11-26 19:10:36.243062] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:23.809 [2024-11-26 19:10:36.243071] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:23.809 [2024-11-26 19:10:36.243078] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:23.809 [2024-11-26 19:10:36.243084] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:23.809 [2024-11-26 19:10:36.243946] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@868 -- # return 0 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:24.381 [2024-11-26 19:10:36.946127] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:24.381 [2024-11-26 19:10:36.970443] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:24.381 NULL1 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@19 -- # rpc_cmd bdev_wait_for_examine 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.381 19:10:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:24.641 19:10:37 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.641 19:10:37 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/fused_ordering/fused_ordering -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:14:24.641 [2024-11-26 19:10:37.041995] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:14:24.641 [2024-11-26 19:10:37.042044] [ DPDK EAL parameters: fused_ordering --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3789954 ] 00:14:25.213 Attached to nqn.2016-06.io.spdk:cnode1 00:14:25.213 Namespace ID: 1 size: 1GB 00:14:25.213 fused_ordering(0) 00:14:25.213 fused_ordering(1) 00:14:25.213 fused_ordering(2) 00:14:25.213 fused_ordering(3) 00:14:25.213 fused_ordering(4) 00:14:25.213 fused_ordering(5) 00:14:25.213 fused_ordering(6) 00:14:25.213 fused_ordering(7) 00:14:25.213 fused_ordering(8) 00:14:25.213 fused_ordering(9) 00:14:25.213 fused_ordering(10) 00:14:25.213 fused_ordering(11) 00:14:25.213 fused_ordering(12) 00:14:25.213 fused_ordering(13) 00:14:25.213 fused_ordering(14) 00:14:25.213 fused_ordering(15) 00:14:25.213 fused_ordering(16) 00:14:25.213 fused_ordering(17) 00:14:25.213 fused_ordering(18) 00:14:25.213 fused_ordering(19) 00:14:25.213 fused_ordering(20) 00:14:25.213 fused_ordering(21) 00:14:25.213 fused_ordering(22) 00:14:25.213 fused_ordering(23) 00:14:25.213 fused_ordering(24) 00:14:25.213 fused_ordering(25) 00:14:25.213 fused_ordering(26) 00:14:25.213 fused_ordering(27) 00:14:25.213 fused_ordering(28) 00:14:25.213 fused_ordering(29) 00:14:25.213 fused_ordering(30) 00:14:25.213 fused_ordering(31) 00:14:25.213 fused_ordering(32) 00:14:25.213 fused_ordering(33) 00:14:25.213 fused_ordering(34) 00:14:25.213 fused_ordering(35) 00:14:25.213 fused_ordering(36) 00:14:25.213 fused_ordering(37) 00:14:25.213 fused_ordering(38) 00:14:25.213 fused_ordering(39) 00:14:25.213 fused_ordering(40) 00:14:25.213 fused_ordering(41) 00:14:25.213 fused_ordering(42) 00:14:25.213 fused_ordering(43) 00:14:25.213 fused_ordering(44) 00:14:25.213 fused_ordering(45) 00:14:25.213 fused_ordering(46) 00:14:25.213 fused_ordering(47) 00:14:25.213 fused_ordering(48) 00:14:25.213 fused_ordering(49) 00:14:25.213 fused_ordering(50) 00:14:25.213 fused_ordering(51) 00:14:25.213 fused_ordering(52) 00:14:25.213 fused_ordering(53) 00:14:25.213 fused_ordering(54) 00:14:25.213 fused_ordering(55) 00:14:25.213 fused_ordering(56) 00:14:25.213 fused_ordering(57) 00:14:25.213 fused_ordering(58) 00:14:25.213 fused_ordering(59) 00:14:25.213 fused_ordering(60) 00:14:25.213 fused_ordering(61) 00:14:25.213 fused_ordering(62) 00:14:25.213 fused_ordering(63) 00:14:25.213 fused_ordering(64) 00:14:25.213 fused_ordering(65) 00:14:25.213 fused_ordering(66) 00:14:25.213 fused_ordering(67) 00:14:25.213 fused_ordering(68) 00:14:25.213 fused_ordering(69) 00:14:25.213 fused_ordering(70) 00:14:25.213 fused_ordering(71) 00:14:25.213 fused_ordering(72) 00:14:25.213 fused_ordering(73) 00:14:25.213 fused_ordering(74) 00:14:25.213 fused_ordering(75) 00:14:25.213 fused_ordering(76) 00:14:25.213 fused_ordering(77) 00:14:25.213 fused_ordering(78) 00:14:25.213 fused_ordering(79) 00:14:25.213 fused_ordering(80) 00:14:25.213 fused_ordering(81) 00:14:25.213 fused_ordering(82) 00:14:25.213 fused_ordering(83) 00:14:25.213 fused_ordering(84) 00:14:25.213 fused_ordering(85) 00:14:25.213 fused_ordering(86) 00:14:25.213 fused_ordering(87) 00:14:25.213 fused_ordering(88) 00:14:25.213 fused_ordering(89) 00:14:25.213 fused_ordering(90) 00:14:25.213 fused_ordering(91) 00:14:25.213 fused_ordering(92) 00:14:25.213 fused_ordering(93) 00:14:25.213 fused_ordering(94) 00:14:25.213 fused_ordering(95) 00:14:25.213 fused_ordering(96) 00:14:25.213 fused_ordering(97) 00:14:25.213 fused_ordering(98) 00:14:25.213 fused_ordering(99) 00:14:25.213 fused_ordering(100) 00:14:25.213 fused_ordering(101) 00:14:25.213 fused_ordering(102) 00:14:25.213 fused_ordering(103) 00:14:25.213 fused_ordering(104) 00:14:25.213 fused_ordering(105) 00:14:25.213 fused_ordering(106) 00:14:25.213 fused_ordering(107) 00:14:25.213 fused_ordering(108) 00:14:25.213 fused_ordering(109) 00:14:25.213 fused_ordering(110) 00:14:25.213 fused_ordering(111) 00:14:25.213 fused_ordering(112) 00:14:25.213 fused_ordering(113) 00:14:25.213 fused_ordering(114) 00:14:25.213 fused_ordering(115) 00:14:25.213 fused_ordering(116) 00:14:25.213 fused_ordering(117) 00:14:25.213 fused_ordering(118) 00:14:25.213 fused_ordering(119) 00:14:25.213 fused_ordering(120) 00:14:25.213 fused_ordering(121) 00:14:25.213 fused_ordering(122) 00:14:25.213 fused_ordering(123) 00:14:25.213 fused_ordering(124) 00:14:25.213 fused_ordering(125) 00:14:25.213 fused_ordering(126) 00:14:25.213 fused_ordering(127) 00:14:25.213 fused_ordering(128) 00:14:25.213 fused_ordering(129) 00:14:25.213 fused_ordering(130) 00:14:25.213 fused_ordering(131) 00:14:25.213 fused_ordering(132) 00:14:25.213 fused_ordering(133) 00:14:25.213 fused_ordering(134) 00:14:25.213 fused_ordering(135) 00:14:25.213 fused_ordering(136) 00:14:25.213 fused_ordering(137) 00:14:25.213 fused_ordering(138) 00:14:25.213 fused_ordering(139) 00:14:25.213 fused_ordering(140) 00:14:25.213 fused_ordering(141) 00:14:25.213 fused_ordering(142) 00:14:25.213 fused_ordering(143) 00:14:25.213 fused_ordering(144) 00:14:25.213 fused_ordering(145) 00:14:25.213 fused_ordering(146) 00:14:25.213 fused_ordering(147) 00:14:25.213 fused_ordering(148) 00:14:25.213 fused_ordering(149) 00:14:25.213 fused_ordering(150) 00:14:25.213 fused_ordering(151) 00:14:25.213 fused_ordering(152) 00:14:25.213 fused_ordering(153) 00:14:25.213 fused_ordering(154) 00:14:25.213 fused_ordering(155) 00:14:25.213 fused_ordering(156) 00:14:25.213 fused_ordering(157) 00:14:25.213 fused_ordering(158) 00:14:25.213 fused_ordering(159) 00:14:25.213 fused_ordering(160) 00:14:25.213 fused_ordering(161) 00:14:25.213 fused_ordering(162) 00:14:25.213 fused_ordering(163) 00:14:25.213 fused_ordering(164) 00:14:25.213 fused_ordering(165) 00:14:25.213 fused_ordering(166) 00:14:25.213 fused_ordering(167) 00:14:25.213 fused_ordering(168) 00:14:25.213 fused_ordering(169) 00:14:25.213 fused_ordering(170) 00:14:25.213 fused_ordering(171) 00:14:25.213 fused_ordering(172) 00:14:25.213 fused_ordering(173) 00:14:25.213 fused_ordering(174) 00:14:25.213 fused_ordering(175) 00:14:25.213 fused_ordering(176) 00:14:25.213 fused_ordering(177) 00:14:25.213 fused_ordering(178) 00:14:25.213 fused_ordering(179) 00:14:25.213 fused_ordering(180) 00:14:25.213 fused_ordering(181) 00:14:25.213 fused_ordering(182) 00:14:25.213 fused_ordering(183) 00:14:25.213 fused_ordering(184) 00:14:25.213 fused_ordering(185) 00:14:25.213 fused_ordering(186) 00:14:25.213 fused_ordering(187) 00:14:25.213 fused_ordering(188) 00:14:25.213 fused_ordering(189) 00:14:25.213 fused_ordering(190) 00:14:25.213 fused_ordering(191) 00:14:25.213 fused_ordering(192) 00:14:25.213 fused_ordering(193) 00:14:25.213 fused_ordering(194) 00:14:25.213 fused_ordering(195) 00:14:25.213 fused_ordering(196) 00:14:25.213 fused_ordering(197) 00:14:25.213 fused_ordering(198) 00:14:25.213 fused_ordering(199) 00:14:25.213 fused_ordering(200) 00:14:25.213 fused_ordering(201) 00:14:25.213 fused_ordering(202) 00:14:25.213 fused_ordering(203) 00:14:25.213 fused_ordering(204) 00:14:25.213 fused_ordering(205) 00:14:25.474 fused_ordering(206) 00:14:25.474 fused_ordering(207) 00:14:25.474 fused_ordering(208) 00:14:25.474 fused_ordering(209) 00:14:25.474 fused_ordering(210) 00:14:25.474 fused_ordering(211) 00:14:25.474 fused_ordering(212) 00:14:25.474 fused_ordering(213) 00:14:25.474 fused_ordering(214) 00:14:25.474 fused_ordering(215) 00:14:25.474 fused_ordering(216) 00:14:25.474 fused_ordering(217) 00:14:25.474 fused_ordering(218) 00:14:25.474 fused_ordering(219) 00:14:25.474 fused_ordering(220) 00:14:25.474 fused_ordering(221) 00:14:25.474 fused_ordering(222) 00:14:25.474 fused_ordering(223) 00:14:25.474 fused_ordering(224) 00:14:25.474 fused_ordering(225) 00:14:25.474 fused_ordering(226) 00:14:25.474 fused_ordering(227) 00:14:25.474 fused_ordering(228) 00:14:25.474 fused_ordering(229) 00:14:25.474 fused_ordering(230) 00:14:25.474 fused_ordering(231) 00:14:25.474 fused_ordering(232) 00:14:25.474 fused_ordering(233) 00:14:25.474 fused_ordering(234) 00:14:25.474 fused_ordering(235) 00:14:25.474 fused_ordering(236) 00:14:25.474 fused_ordering(237) 00:14:25.474 fused_ordering(238) 00:14:25.474 fused_ordering(239) 00:14:25.474 fused_ordering(240) 00:14:25.474 fused_ordering(241) 00:14:25.474 fused_ordering(242) 00:14:25.474 fused_ordering(243) 00:14:25.474 fused_ordering(244) 00:14:25.474 fused_ordering(245) 00:14:25.474 fused_ordering(246) 00:14:25.474 fused_ordering(247) 00:14:25.474 fused_ordering(248) 00:14:25.474 fused_ordering(249) 00:14:25.474 fused_ordering(250) 00:14:25.474 fused_ordering(251) 00:14:25.474 fused_ordering(252) 00:14:25.474 fused_ordering(253) 00:14:25.474 fused_ordering(254) 00:14:25.474 fused_ordering(255) 00:14:25.474 fused_ordering(256) 00:14:25.474 fused_ordering(257) 00:14:25.474 fused_ordering(258) 00:14:25.474 fused_ordering(259) 00:14:25.474 fused_ordering(260) 00:14:25.474 fused_ordering(261) 00:14:25.474 fused_ordering(262) 00:14:25.474 fused_ordering(263) 00:14:25.474 fused_ordering(264) 00:14:25.474 fused_ordering(265) 00:14:25.474 fused_ordering(266) 00:14:25.474 fused_ordering(267) 00:14:25.474 fused_ordering(268) 00:14:25.474 fused_ordering(269) 00:14:25.474 fused_ordering(270) 00:14:25.474 fused_ordering(271) 00:14:25.474 fused_ordering(272) 00:14:25.474 fused_ordering(273) 00:14:25.474 fused_ordering(274) 00:14:25.474 fused_ordering(275) 00:14:25.474 fused_ordering(276) 00:14:25.474 fused_ordering(277) 00:14:25.474 fused_ordering(278) 00:14:25.474 fused_ordering(279) 00:14:25.474 fused_ordering(280) 00:14:25.474 fused_ordering(281) 00:14:25.474 fused_ordering(282) 00:14:25.474 fused_ordering(283) 00:14:25.474 fused_ordering(284) 00:14:25.474 fused_ordering(285) 00:14:25.474 fused_ordering(286) 00:14:25.474 fused_ordering(287) 00:14:25.474 fused_ordering(288) 00:14:25.474 fused_ordering(289) 00:14:25.475 fused_ordering(290) 00:14:25.475 fused_ordering(291) 00:14:25.475 fused_ordering(292) 00:14:25.475 fused_ordering(293) 00:14:25.475 fused_ordering(294) 00:14:25.475 fused_ordering(295) 00:14:25.475 fused_ordering(296) 00:14:25.475 fused_ordering(297) 00:14:25.475 fused_ordering(298) 00:14:25.475 fused_ordering(299) 00:14:25.475 fused_ordering(300) 00:14:25.475 fused_ordering(301) 00:14:25.475 fused_ordering(302) 00:14:25.475 fused_ordering(303) 00:14:25.475 fused_ordering(304) 00:14:25.475 fused_ordering(305) 00:14:25.475 fused_ordering(306) 00:14:25.475 fused_ordering(307) 00:14:25.475 fused_ordering(308) 00:14:25.475 fused_ordering(309) 00:14:25.475 fused_ordering(310) 00:14:25.475 fused_ordering(311) 00:14:25.475 fused_ordering(312) 00:14:25.475 fused_ordering(313) 00:14:25.475 fused_ordering(314) 00:14:25.475 fused_ordering(315) 00:14:25.475 fused_ordering(316) 00:14:25.475 fused_ordering(317) 00:14:25.475 fused_ordering(318) 00:14:25.475 fused_ordering(319) 00:14:25.475 fused_ordering(320) 00:14:25.475 fused_ordering(321) 00:14:25.475 fused_ordering(322) 00:14:25.475 fused_ordering(323) 00:14:25.475 fused_ordering(324) 00:14:25.475 fused_ordering(325) 00:14:25.475 fused_ordering(326) 00:14:25.475 fused_ordering(327) 00:14:25.475 fused_ordering(328) 00:14:25.475 fused_ordering(329) 00:14:25.475 fused_ordering(330) 00:14:25.475 fused_ordering(331) 00:14:25.475 fused_ordering(332) 00:14:25.475 fused_ordering(333) 00:14:25.475 fused_ordering(334) 00:14:25.475 fused_ordering(335) 00:14:25.475 fused_ordering(336) 00:14:25.475 fused_ordering(337) 00:14:25.475 fused_ordering(338) 00:14:25.475 fused_ordering(339) 00:14:25.475 fused_ordering(340) 00:14:25.475 fused_ordering(341) 00:14:25.475 fused_ordering(342) 00:14:25.475 fused_ordering(343) 00:14:25.475 fused_ordering(344) 00:14:25.475 fused_ordering(345) 00:14:25.475 fused_ordering(346) 00:14:25.475 fused_ordering(347) 00:14:25.475 fused_ordering(348) 00:14:25.475 fused_ordering(349) 00:14:25.475 fused_ordering(350) 00:14:25.475 fused_ordering(351) 00:14:25.475 fused_ordering(352) 00:14:25.475 fused_ordering(353) 00:14:25.475 fused_ordering(354) 00:14:25.475 fused_ordering(355) 00:14:25.475 fused_ordering(356) 00:14:25.475 fused_ordering(357) 00:14:25.475 fused_ordering(358) 00:14:25.475 fused_ordering(359) 00:14:25.475 fused_ordering(360) 00:14:25.475 fused_ordering(361) 00:14:25.475 fused_ordering(362) 00:14:25.475 fused_ordering(363) 00:14:25.475 fused_ordering(364) 00:14:25.475 fused_ordering(365) 00:14:25.475 fused_ordering(366) 00:14:25.475 fused_ordering(367) 00:14:25.475 fused_ordering(368) 00:14:25.475 fused_ordering(369) 00:14:25.475 fused_ordering(370) 00:14:25.475 fused_ordering(371) 00:14:25.475 fused_ordering(372) 00:14:25.475 fused_ordering(373) 00:14:25.475 fused_ordering(374) 00:14:25.475 fused_ordering(375) 00:14:25.475 fused_ordering(376) 00:14:25.475 fused_ordering(377) 00:14:25.475 fused_ordering(378) 00:14:25.475 fused_ordering(379) 00:14:25.475 fused_ordering(380) 00:14:25.475 fused_ordering(381) 00:14:25.475 fused_ordering(382) 00:14:25.475 fused_ordering(383) 00:14:25.475 fused_ordering(384) 00:14:25.475 fused_ordering(385) 00:14:25.475 fused_ordering(386) 00:14:25.475 fused_ordering(387) 00:14:25.475 fused_ordering(388) 00:14:25.475 fused_ordering(389) 00:14:25.475 fused_ordering(390) 00:14:25.475 fused_ordering(391) 00:14:25.475 fused_ordering(392) 00:14:25.475 fused_ordering(393) 00:14:25.475 fused_ordering(394) 00:14:25.475 fused_ordering(395) 00:14:25.475 fused_ordering(396) 00:14:25.475 fused_ordering(397) 00:14:25.475 fused_ordering(398) 00:14:25.475 fused_ordering(399) 00:14:25.475 fused_ordering(400) 00:14:25.475 fused_ordering(401) 00:14:25.475 fused_ordering(402) 00:14:25.475 fused_ordering(403) 00:14:25.475 fused_ordering(404) 00:14:25.475 fused_ordering(405) 00:14:25.475 fused_ordering(406) 00:14:25.475 fused_ordering(407) 00:14:25.475 fused_ordering(408) 00:14:25.475 fused_ordering(409) 00:14:25.475 fused_ordering(410) 00:14:25.735 fused_ordering(411) 00:14:25.735 fused_ordering(412) 00:14:25.735 fused_ordering(413) 00:14:25.735 fused_ordering(414) 00:14:25.735 fused_ordering(415) 00:14:25.735 fused_ordering(416) 00:14:25.735 fused_ordering(417) 00:14:25.735 fused_ordering(418) 00:14:25.735 fused_ordering(419) 00:14:25.735 fused_ordering(420) 00:14:25.735 fused_ordering(421) 00:14:25.735 fused_ordering(422) 00:14:25.735 fused_ordering(423) 00:14:25.735 fused_ordering(424) 00:14:25.735 fused_ordering(425) 00:14:25.735 fused_ordering(426) 00:14:25.735 fused_ordering(427) 00:14:25.735 fused_ordering(428) 00:14:25.735 fused_ordering(429) 00:14:25.735 fused_ordering(430) 00:14:25.735 fused_ordering(431) 00:14:25.735 fused_ordering(432) 00:14:25.735 fused_ordering(433) 00:14:25.735 fused_ordering(434) 00:14:25.735 fused_ordering(435) 00:14:25.735 fused_ordering(436) 00:14:25.735 fused_ordering(437) 00:14:25.735 fused_ordering(438) 00:14:25.735 fused_ordering(439) 00:14:25.735 fused_ordering(440) 00:14:25.735 fused_ordering(441) 00:14:25.735 fused_ordering(442) 00:14:25.735 fused_ordering(443) 00:14:25.735 fused_ordering(444) 00:14:25.735 fused_ordering(445) 00:14:25.735 fused_ordering(446) 00:14:25.735 fused_ordering(447) 00:14:25.735 fused_ordering(448) 00:14:25.735 fused_ordering(449) 00:14:25.735 fused_ordering(450) 00:14:25.735 fused_ordering(451) 00:14:25.735 fused_ordering(452) 00:14:25.735 fused_ordering(453) 00:14:25.735 fused_ordering(454) 00:14:25.735 fused_ordering(455) 00:14:25.735 fused_ordering(456) 00:14:25.735 fused_ordering(457) 00:14:25.735 fused_ordering(458) 00:14:25.735 fused_ordering(459) 00:14:25.736 fused_ordering(460) 00:14:25.736 fused_ordering(461) 00:14:25.736 fused_ordering(462) 00:14:25.736 fused_ordering(463) 00:14:25.736 fused_ordering(464) 00:14:25.736 fused_ordering(465) 00:14:25.736 fused_ordering(466) 00:14:25.736 fused_ordering(467) 00:14:25.736 fused_ordering(468) 00:14:25.736 fused_ordering(469) 00:14:25.736 fused_ordering(470) 00:14:25.736 fused_ordering(471) 00:14:25.736 fused_ordering(472) 00:14:25.736 fused_ordering(473) 00:14:25.736 fused_ordering(474) 00:14:25.736 fused_ordering(475) 00:14:25.736 fused_ordering(476) 00:14:25.736 fused_ordering(477) 00:14:25.736 fused_ordering(478) 00:14:25.736 fused_ordering(479) 00:14:25.736 fused_ordering(480) 00:14:25.736 fused_ordering(481) 00:14:25.736 fused_ordering(482) 00:14:25.736 fused_ordering(483) 00:14:25.736 fused_ordering(484) 00:14:25.736 fused_ordering(485) 00:14:25.736 fused_ordering(486) 00:14:25.736 fused_ordering(487) 00:14:25.736 fused_ordering(488) 00:14:25.736 fused_ordering(489) 00:14:25.736 fused_ordering(490) 00:14:25.736 fused_ordering(491) 00:14:25.736 fused_ordering(492) 00:14:25.736 fused_ordering(493) 00:14:25.736 fused_ordering(494) 00:14:25.736 fused_ordering(495) 00:14:25.736 fused_ordering(496) 00:14:25.736 fused_ordering(497) 00:14:25.736 fused_ordering(498) 00:14:25.736 fused_ordering(499) 00:14:25.736 fused_ordering(500) 00:14:25.736 fused_ordering(501) 00:14:25.736 fused_ordering(502) 00:14:25.736 fused_ordering(503) 00:14:25.736 fused_ordering(504) 00:14:25.736 fused_ordering(505) 00:14:25.736 fused_ordering(506) 00:14:25.736 fused_ordering(507) 00:14:25.736 fused_ordering(508) 00:14:25.736 fused_ordering(509) 00:14:25.736 fused_ordering(510) 00:14:25.736 fused_ordering(511) 00:14:25.736 fused_ordering(512) 00:14:25.736 fused_ordering(513) 00:14:25.736 fused_ordering(514) 00:14:25.736 fused_ordering(515) 00:14:25.736 fused_ordering(516) 00:14:25.736 fused_ordering(517) 00:14:25.736 fused_ordering(518) 00:14:25.736 fused_ordering(519) 00:14:25.736 fused_ordering(520) 00:14:25.736 fused_ordering(521) 00:14:25.736 fused_ordering(522) 00:14:25.736 fused_ordering(523) 00:14:25.736 fused_ordering(524) 00:14:25.736 fused_ordering(525) 00:14:25.736 fused_ordering(526) 00:14:25.736 fused_ordering(527) 00:14:25.736 fused_ordering(528) 00:14:25.736 fused_ordering(529) 00:14:25.736 fused_ordering(530) 00:14:25.736 fused_ordering(531) 00:14:25.736 fused_ordering(532) 00:14:25.736 fused_ordering(533) 00:14:25.736 fused_ordering(534) 00:14:25.736 fused_ordering(535) 00:14:25.736 fused_ordering(536) 00:14:25.736 fused_ordering(537) 00:14:25.736 fused_ordering(538) 00:14:25.736 fused_ordering(539) 00:14:25.736 fused_ordering(540) 00:14:25.736 fused_ordering(541) 00:14:25.736 fused_ordering(542) 00:14:25.736 fused_ordering(543) 00:14:25.736 fused_ordering(544) 00:14:25.736 fused_ordering(545) 00:14:25.736 fused_ordering(546) 00:14:25.736 fused_ordering(547) 00:14:25.736 fused_ordering(548) 00:14:25.736 fused_ordering(549) 00:14:25.736 fused_ordering(550) 00:14:25.736 fused_ordering(551) 00:14:25.736 fused_ordering(552) 00:14:25.736 fused_ordering(553) 00:14:25.736 fused_ordering(554) 00:14:25.736 fused_ordering(555) 00:14:25.736 fused_ordering(556) 00:14:25.736 fused_ordering(557) 00:14:25.736 fused_ordering(558) 00:14:25.736 fused_ordering(559) 00:14:25.736 fused_ordering(560) 00:14:25.736 fused_ordering(561) 00:14:25.736 fused_ordering(562) 00:14:25.736 fused_ordering(563) 00:14:25.736 fused_ordering(564) 00:14:25.736 fused_ordering(565) 00:14:25.736 fused_ordering(566) 00:14:25.736 fused_ordering(567) 00:14:25.736 fused_ordering(568) 00:14:25.736 fused_ordering(569) 00:14:25.736 fused_ordering(570) 00:14:25.736 fused_ordering(571) 00:14:25.736 fused_ordering(572) 00:14:25.736 fused_ordering(573) 00:14:25.736 fused_ordering(574) 00:14:25.736 fused_ordering(575) 00:14:25.736 fused_ordering(576) 00:14:25.736 fused_ordering(577) 00:14:25.736 fused_ordering(578) 00:14:25.736 fused_ordering(579) 00:14:25.736 fused_ordering(580) 00:14:25.736 fused_ordering(581) 00:14:25.736 fused_ordering(582) 00:14:25.736 fused_ordering(583) 00:14:25.736 fused_ordering(584) 00:14:25.736 fused_ordering(585) 00:14:25.736 fused_ordering(586) 00:14:25.736 fused_ordering(587) 00:14:25.736 fused_ordering(588) 00:14:25.736 fused_ordering(589) 00:14:25.736 fused_ordering(590) 00:14:25.736 fused_ordering(591) 00:14:25.736 fused_ordering(592) 00:14:25.736 fused_ordering(593) 00:14:25.736 fused_ordering(594) 00:14:25.736 fused_ordering(595) 00:14:25.736 fused_ordering(596) 00:14:25.736 fused_ordering(597) 00:14:25.736 fused_ordering(598) 00:14:25.736 fused_ordering(599) 00:14:25.736 fused_ordering(600) 00:14:25.736 fused_ordering(601) 00:14:25.736 fused_ordering(602) 00:14:25.736 fused_ordering(603) 00:14:25.736 fused_ordering(604) 00:14:25.736 fused_ordering(605) 00:14:25.736 fused_ordering(606) 00:14:25.736 fused_ordering(607) 00:14:25.736 fused_ordering(608) 00:14:25.736 fused_ordering(609) 00:14:25.736 fused_ordering(610) 00:14:25.736 fused_ordering(611) 00:14:25.736 fused_ordering(612) 00:14:25.736 fused_ordering(613) 00:14:25.736 fused_ordering(614) 00:14:25.736 fused_ordering(615) 00:14:26.306 fused_ordering(616) 00:14:26.306 fused_ordering(617) 00:14:26.306 fused_ordering(618) 00:14:26.306 fused_ordering(619) 00:14:26.306 fused_ordering(620) 00:14:26.306 fused_ordering(621) 00:14:26.306 fused_ordering(622) 00:14:26.306 fused_ordering(623) 00:14:26.306 fused_ordering(624) 00:14:26.306 fused_ordering(625) 00:14:26.306 fused_ordering(626) 00:14:26.306 fused_ordering(627) 00:14:26.306 fused_ordering(628) 00:14:26.306 fused_ordering(629) 00:14:26.306 fused_ordering(630) 00:14:26.306 fused_ordering(631) 00:14:26.306 fused_ordering(632) 00:14:26.306 fused_ordering(633) 00:14:26.306 fused_ordering(634) 00:14:26.306 fused_ordering(635) 00:14:26.306 fused_ordering(636) 00:14:26.306 fused_ordering(637) 00:14:26.306 fused_ordering(638) 00:14:26.306 fused_ordering(639) 00:14:26.306 fused_ordering(640) 00:14:26.306 fused_ordering(641) 00:14:26.306 fused_ordering(642) 00:14:26.306 fused_ordering(643) 00:14:26.306 fused_ordering(644) 00:14:26.306 fused_ordering(645) 00:14:26.306 fused_ordering(646) 00:14:26.306 fused_ordering(647) 00:14:26.306 fused_ordering(648) 00:14:26.306 fused_ordering(649) 00:14:26.306 fused_ordering(650) 00:14:26.306 fused_ordering(651) 00:14:26.306 fused_ordering(652) 00:14:26.306 fused_ordering(653) 00:14:26.306 fused_ordering(654) 00:14:26.306 fused_ordering(655) 00:14:26.306 fused_ordering(656) 00:14:26.306 fused_ordering(657) 00:14:26.306 fused_ordering(658) 00:14:26.306 fused_ordering(659) 00:14:26.306 fused_ordering(660) 00:14:26.306 fused_ordering(661) 00:14:26.306 fused_ordering(662) 00:14:26.306 fused_ordering(663) 00:14:26.306 fused_ordering(664) 00:14:26.306 fused_ordering(665) 00:14:26.306 fused_ordering(666) 00:14:26.306 fused_ordering(667) 00:14:26.306 fused_ordering(668) 00:14:26.306 fused_ordering(669) 00:14:26.306 fused_ordering(670) 00:14:26.306 fused_ordering(671) 00:14:26.306 fused_ordering(672) 00:14:26.306 fused_ordering(673) 00:14:26.306 fused_ordering(674) 00:14:26.306 fused_ordering(675) 00:14:26.306 fused_ordering(676) 00:14:26.306 fused_ordering(677) 00:14:26.306 fused_ordering(678) 00:14:26.306 fused_ordering(679) 00:14:26.306 fused_ordering(680) 00:14:26.306 fused_ordering(681) 00:14:26.306 fused_ordering(682) 00:14:26.306 fused_ordering(683) 00:14:26.306 fused_ordering(684) 00:14:26.306 fused_ordering(685) 00:14:26.306 fused_ordering(686) 00:14:26.306 fused_ordering(687) 00:14:26.306 fused_ordering(688) 00:14:26.306 fused_ordering(689) 00:14:26.306 fused_ordering(690) 00:14:26.306 fused_ordering(691) 00:14:26.306 fused_ordering(692) 00:14:26.306 fused_ordering(693) 00:14:26.306 fused_ordering(694) 00:14:26.306 fused_ordering(695) 00:14:26.306 fused_ordering(696) 00:14:26.306 fused_ordering(697) 00:14:26.306 fused_ordering(698) 00:14:26.306 fused_ordering(699) 00:14:26.306 fused_ordering(700) 00:14:26.306 fused_ordering(701) 00:14:26.306 fused_ordering(702) 00:14:26.306 fused_ordering(703) 00:14:26.306 fused_ordering(704) 00:14:26.306 fused_ordering(705) 00:14:26.306 fused_ordering(706) 00:14:26.306 fused_ordering(707) 00:14:26.306 fused_ordering(708) 00:14:26.306 fused_ordering(709) 00:14:26.306 fused_ordering(710) 00:14:26.306 fused_ordering(711) 00:14:26.306 fused_ordering(712) 00:14:26.306 fused_ordering(713) 00:14:26.306 fused_ordering(714) 00:14:26.306 fused_ordering(715) 00:14:26.306 fused_ordering(716) 00:14:26.306 fused_ordering(717) 00:14:26.306 fused_ordering(718) 00:14:26.306 fused_ordering(719) 00:14:26.306 fused_ordering(720) 00:14:26.306 fused_ordering(721) 00:14:26.306 fused_ordering(722) 00:14:26.306 fused_ordering(723) 00:14:26.306 fused_ordering(724) 00:14:26.306 fused_ordering(725) 00:14:26.306 fused_ordering(726) 00:14:26.306 fused_ordering(727) 00:14:26.306 fused_ordering(728) 00:14:26.306 fused_ordering(729) 00:14:26.306 fused_ordering(730) 00:14:26.306 fused_ordering(731) 00:14:26.306 fused_ordering(732) 00:14:26.306 fused_ordering(733) 00:14:26.306 fused_ordering(734) 00:14:26.306 fused_ordering(735) 00:14:26.306 fused_ordering(736) 00:14:26.306 fused_ordering(737) 00:14:26.306 fused_ordering(738) 00:14:26.306 fused_ordering(739) 00:14:26.306 fused_ordering(740) 00:14:26.306 fused_ordering(741) 00:14:26.306 fused_ordering(742) 00:14:26.306 fused_ordering(743) 00:14:26.306 fused_ordering(744) 00:14:26.306 fused_ordering(745) 00:14:26.306 fused_ordering(746) 00:14:26.306 fused_ordering(747) 00:14:26.306 fused_ordering(748) 00:14:26.306 fused_ordering(749) 00:14:26.306 fused_ordering(750) 00:14:26.306 fused_ordering(751) 00:14:26.306 fused_ordering(752) 00:14:26.306 fused_ordering(753) 00:14:26.306 fused_ordering(754) 00:14:26.306 fused_ordering(755) 00:14:26.306 fused_ordering(756) 00:14:26.306 fused_ordering(757) 00:14:26.306 fused_ordering(758) 00:14:26.306 fused_ordering(759) 00:14:26.306 fused_ordering(760) 00:14:26.306 fused_ordering(761) 00:14:26.306 fused_ordering(762) 00:14:26.306 fused_ordering(763) 00:14:26.306 fused_ordering(764) 00:14:26.306 fused_ordering(765) 00:14:26.306 fused_ordering(766) 00:14:26.306 fused_ordering(767) 00:14:26.306 fused_ordering(768) 00:14:26.306 fused_ordering(769) 00:14:26.306 fused_ordering(770) 00:14:26.306 fused_ordering(771) 00:14:26.306 fused_ordering(772) 00:14:26.306 fused_ordering(773) 00:14:26.306 fused_ordering(774) 00:14:26.306 fused_ordering(775) 00:14:26.306 fused_ordering(776) 00:14:26.306 fused_ordering(777) 00:14:26.306 fused_ordering(778) 00:14:26.306 fused_ordering(779) 00:14:26.306 fused_ordering(780) 00:14:26.306 fused_ordering(781) 00:14:26.306 fused_ordering(782) 00:14:26.306 fused_ordering(783) 00:14:26.306 fused_ordering(784) 00:14:26.306 fused_ordering(785) 00:14:26.306 fused_ordering(786) 00:14:26.306 fused_ordering(787) 00:14:26.306 fused_ordering(788) 00:14:26.306 fused_ordering(789) 00:14:26.306 fused_ordering(790) 00:14:26.306 fused_ordering(791) 00:14:26.306 fused_ordering(792) 00:14:26.306 fused_ordering(793) 00:14:26.306 fused_ordering(794) 00:14:26.306 fused_ordering(795) 00:14:26.306 fused_ordering(796) 00:14:26.306 fused_ordering(797) 00:14:26.306 fused_ordering(798) 00:14:26.306 fused_ordering(799) 00:14:26.306 fused_ordering(800) 00:14:26.306 fused_ordering(801) 00:14:26.306 fused_ordering(802) 00:14:26.306 fused_ordering(803) 00:14:26.306 fused_ordering(804) 00:14:26.306 fused_ordering(805) 00:14:26.306 fused_ordering(806) 00:14:26.306 fused_ordering(807) 00:14:26.306 fused_ordering(808) 00:14:26.306 fused_ordering(809) 00:14:26.306 fused_ordering(810) 00:14:26.306 fused_ordering(811) 00:14:26.306 fused_ordering(812) 00:14:26.306 fused_ordering(813) 00:14:26.306 fused_ordering(814) 00:14:26.306 fused_ordering(815) 00:14:26.306 fused_ordering(816) 00:14:26.306 fused_ordering(817) 00:14:26.306 fused_ordering(818) 00:14:26.306 fused_ordering(819) 00:14:26.306 fused_ordering(820) 00:14:26.875 fused_ordering(821) 00:14:26.875 fused_ordering(822) 00:14:26.875 fused_ordering(823) 00:14:26.875 fused_ordering(824) 00:14:26.875 fused_ordering(825) 00:14:26.875 fused_ordering(826) 00:14:26.875 fused_ordering(827) 00:14:26.875 fused_ordering(828) 00:14:26.875 fused_ordering(829) 00:14:26.875 fused_ordering(830) 00:14:26.875 fused_ordering(831) 00:14:26.875 fused_ordering(832) 00:14:26.875 fused_ordering(833) 00:14:26.875 fused_ordering(834) 00:14:26.875 fused_ordering(835) 00:14:26.875 fused_ordering(836) 00:14:26.875 fused_ordering(837) 00:14:26.875 fused_ordering(838) 00:14:26.875 fused_ordering(839) 00:14:26.875 fused_ordering(840) 00:14:26.875 fused_ordering(841) 00:14:26.875 fused_ordering(842) 00:14:26.875 fused_ordering(843) 00:14:26.875 fused_ordering(844) 00:14:26.875 fused_ordering(845) 00:14:26.875 fused_ordering(846) 00:14:26.875 fused_ordering(847) 00:14:26.875 fused_ordering(848) 00:14:26.875 fused_ordering(849) 00:14:26.875 fused_ordering(850) 00:14:26.875 fused_ordering(851) 00:14:26.875 fused_ordering(852) 00:14:26.875 fused_ordering(853) 00:14:26.875 fused_ordering(854) 00:14:26.875 fused_ordering(855) 00:14:26.875 fused_ordering(856) 00:14:26.875 fused_ordering(857) 00:14:26.875 fused_ordering(858) 00:14:26.875 fused_ordering(859) 00:14:26.875 fused_ordering(860) 00:14:26.875 fused_ordering(861) 00:14:26.875 fused_ordering(862) 00:14:26.875 fused_ordering(863) 00:14:26.875 fused_ordering(864) 00:14:26.875 fused_ordering(865) 00:14:26.875 fused_ordering(866) 00:14:26.875 fused_ordering(867) 00:14:26.875 fused_ordering(868) 00:14:26.875 fused_ordering(869) 00:14:26.875 fused_ordering(870) 00:14:26.875 fused_ordering(871) 00:14:26.875 fused_ordering(872) 00:14:26.875 fused_ordering(873) 00:14:26.875 fused_ordering(874) 00:14:26.875 fused_ordering(875) 00:14:26.875 fused_ordering(876) 00:14:26.875 fused_ordering(877) 00:14:26.875 fused_ordering(878) 00:14:26.875 fused_ordering(879) 00:14:26.875 fused_ordering(880) 00:14:26.875 fused_ordering(881) 00:14:26.875 fused_ordering(882) 00:14:26.875 fused_ordering(883) 00:14:26.875 fused_ordering(884) 00:14:26.875 fused_ordering(885) 00:14:26.875 fused_ordering(886) 00:14:26.875 fused_ordering(887) 00:14:26.875 fused_ordering(888) 00:14:26.875 fused_ordering(889) 00:14:26.875 fused_ordering(890) 00:14:26.875 fused_ordering(891) 00:14:26.875 fused_ordering(892) 00:14:26.875 fused_ordering(893) 00:14:26.875 fused_ordering(894) 00:14:26.875 fused_ordering(895) 00:14:26.875 fused_ordering(896) 00:14:26.875 fused_ordering(897) 00:14:26.875 fused_ordering(898) 00:14:26.876 fused_ordering(899) 00:14:26.876 fused_ordering(900) 00:14:26.876 fused_ordering(901) 00:14:26.876 fused_ordering(902) 00:14:26.876 fused_ordering(903) 00:14:26.876 fused_ordering(904) 00:14:26.876 fused_ordering(905) 00:14:26.876 fused_ordering(906) 00:14:26.876 fused_ordering(907) 00:14:26.876 fused_ordering(908) 00:14:26.876 fused_ordering(909) 00:14:26.876 fused_ordering(910) 00:14:26.876 fused_ordering(911) 00:14:26.876 fused_ordering(912) 00:14:26.876 fused_ordering(913) 00:14:26.876 fused_ordering(914) 00:14:26.876 fused_ordering(915) 00:14:26.876 fused_ordering(916) 00:14:26.876 fused_ordering(917) 00:14:26.876 fused_ordering(918) 00:14:26.876 fused_ordering(919) 00:14:26.876 fused_ordering(920) 00:14:26.876 fused_ordering(921) 00:14:26.876 fused_ordering(922) 00:14:26.876 fused_ordering(923) 00:14:26.876 fused_ordering(924) 00:14:26.876 fused_ordering(925) 00:14:26.876 fused_ordering(926) 00:14:26.876 fused_ordering(927) 00:14:26.876 fused_ordering(928) 00:14:26.876 fused_ordering(929) 00:14:26.876 fused_ordering(930) 00:14:26.876 fused_ordering(931) 00:14:26.876 fused_ordering(932) 00:14:26.876 fused_ordering(933) 00:14:26.876 fused_ordering(934) 00:14:26.876 fused_ordering(935) 00:14:26.876 fused_ordering(936) 00:14:26.876 fused_ordering(937) 00:14:26.876 fused_ordering(938) 00:14:26.876 fused_ordering(939) 00:14:26.876 fused_ordering(940) 00:14:26.876 fused_ordering(941) 00:14:26.876 fused_ordering(942) 00:14:26.876 fused_ordering(943) 00:14:26.876 fused_ordering(944) 00:14:26.876 fused_ordering(945) 00:14:26.876 fused_ordering(946) 00:14:26.876 fused_ordering(947) 00:14:26.876 fused_ordering(948) 00:14:26.876 fused_ordering(949) 00:14:26.876 fused_ordering(950) 00:14:26.876 fused_ordering(951) 00:14:26.876 fused_ordering(952) 00:14:26.876 fused_ordering(953) 00:14:26.876 fused_ordering(954) 00:14:26.876 fused_ordering(955) 00:14:26.876 fused_ordering(956) 00:14:26.876 fused_ordering(957) 00:14:26.876 fused_ordering(958) 00:14:26.876 fused_ordering(959) 00:14:26.876 fused_ordering(960) 00:14:26.876 fused_ordering(961) 00:14:26.876 fused_ordering(962) 00:14:26.876 fused_ordering(963) 00:14:26.876 fused_ordering(964) 00:14:26.876 fused_ordering(965) 00:14:26.876 fused_ordering(966) 00:14:26.876 fused_ordering(967) 00:14:26.876 fused_ordering(968) 00:14:26.876 fused_ordering(969) 00:14:26.876 fused_ordering(970) 00:14:26.876 fused_ordering(971) 00:14:26.876 fused_ordering(972) 00:14:26.876 fused_ordering(973) 00:14:26.876 fused_ordering(974) 00:14:26.876 fused_ordering(975) 00:14:26.876 fused_ordering(976) 00:14:26.876 fused_ordering(977) 00:14:26.876 fused_ordering(978) 00:14:26.876 fused_ordering(979) 00:14:26.876 fused_ordering(980) 00:14:26.876 fused_ordering(981) 00:14:26.876 fused_ordering(982) 00:14:26.876 fused_ordering(983) 00:14:26.876 fused_ordering(984) 00:14:26.876 fused_ordering(985) 00:14:26.876 fused_ordering(986) 00:14:26.876 fused_ordering(987) 00:14:26.876 fused_ordering(988) 00:14:26.876 fused_ordering(989) 00:14:26.876 fused_ordering(990) 00:14:26.876 fused_ordering(991) 00:14:26.876 fused_ordering(992) 00:14:26.876 fused_ordering(993) 00:14:26.876 fused_ordering(994) 00:14:26.876 fused_ordering(995) 00:14:26.876 fused_ordering(996) 00:14:26.876 fused_ordering(997) 00:14:26.876 fused_ordering(998) 00:14:26.876 fused_ordering(999) 00:14:26.876 fused_ordering(1000) 00:14:26.876 fused_ordering(1001) 00:14:26.876 fused_ordering(1002) 00:14:26.876 fused_ordering(1003) 00:14:26.876 fused_ordering(1004) 00:14:26.876 fused_ordering(1005) 00:14:26.876 fused_ordering(1006) 00:14:26.876 fused_ordering(1007) 00:14:26.876 fused_ordering(1008) 00:14:26.876 fused_ordering(1009) 00:14:26.876 fused_ordering(1010) 00:14:26.876 fused_ordering(1011) 00:14:26.876 fused_ordering(1012) 00:14:26.876 fused_ordering(1013) 00:14:26.876 fused_ordering(1014) 00:14:26.876 fused_ordering(1015) 00:14:26.876 fused_ordering(1016) 00:14:26.876 fused_ordering(1017) 00:14:26.876 fused_ordering(1018) 00:14:26.876 fused_ordering(1019) 00:14:26.876 fused_ordering(1020) 00:14:26.876 fused_ordering(1021) 00:14:26.876 fused_ordering(1022) 00:14:26.876 fused_ordering(1023) 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@23 -- # trap - SIGINT SIGTERM EXIT 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@25 -- # nvmftestfini 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@516 -- # nvmfcleanup 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@121 -- # sync 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@124 -- # set +e 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:26.876 rmmod nvme_tcp 00:14:26.876 rmmod nvme_fabrics 00:14:26.876 rmmod nvme_keyring 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@128 -- # set -e 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@129 -- # return 0 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@517 -- # '[' -n 3789862 ']' 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@518 -- # killprocess 3789862 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@954 -- # '[' -z 3789862 ']' 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@958 -- # kill -0 3789862 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@959 -- # uname 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:26.876 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3789862 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3789862' 00:14:27.135 killing process with pid 3789862 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@973 -- # kill 3789862 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@978 -- # wait 3789862 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@297 -- # iptr 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@791 -- # iptables-save 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@791 -- # iptables-restore 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:27.135 19:10:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:14:29.674 00:14:29.674 real 0m14.440s 00:14:29.674 user 0m7.397s 00:14:29.674 sys 0m7.856s 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:14:29.674 ************************************ 00:14:29.674 END TEST nvmf_fused_ordering 00:14:29.674 ************************************ 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@26 -- # run_test nvmf_ns_masking test/nvmf/target/ns_masking.sh --transport=tcp 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:14:29.674 ************************************ 00:14:29.674 START TEST nvmf_ns_masking 00:14:29.674 ************************************ 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1129 -- # test/nvmf/target/ns_masking.sh --transport=tcp 00:14:29.674 * Looking for test storage... 00:14:29.674 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1693 -- # lcov --version 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # IFS=.-: 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # read -ra ver1 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # IFS=.-: 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # read -ra ver2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@338 -- # local 'op=<' 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@340 -- # ver1_l=2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@341 -- # ver2_l=1 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@344 -- # case "$op" in 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@345 -- # : 1 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # decimal 1 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=1 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 1 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # ver1[v]=1 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # decimal 2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # ver2[v]=2 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:29.674 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # return 0 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:14:29.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:29.675 --rc genhtml_branch_coverage=1 00:14:29.675 --rc genhtml_function_coverage=1 00:14:29.675 --rc genhtml_legend=1 00:14:29.675 --rc geninfo_all_blocks=1 00:14:29.675 --rc geninfo_unexecuted_blocks=1 00:14:29.675 00:14:29.675 ' 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:14:29.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:29.675 --rc genhtml_branch_coverage=1 00:14:29.675 --rc genhtml_function_coverage=1 00:14:29.675 --rc genhtml_legend=1 00:14:29.675 --rc geninfo_all_blocks=1 00:14:29.675 --rc geninfo_unexecuted_blocks=1 00:14:29.675 00:14:29.675 ' 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:14:29.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:29.675 --rc genhtml_branch_coverage=1 00:14:29.675 --rc genhtml_function_coverage=1 00:14:29.675 --rc genhtml_legend=1 00:14:29.675 --rc geninfo_all_blocks=1 00:14:29.675 --rc geninfo_unexecuted_blocks=1 00:14:29.675 00:14:29.675 ' 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:14:29.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:29.675 --rc genhtml_branch_coverage=1 00:14:29.675 --rc genhtml_function_coverage=1 00:14:29.675 --rc genhtml_legend=1 00:14:29.675 --rc geninfo_all_blocks=1 00:14:29.675 --rc geninfo_unexecuted_blocks=1 00:14:29.675 00:14:29.675 ' 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # uname -s 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:29.675 19:10:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@15 -- # shopt -s extglob 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@5 -- # export PATH 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@51 -- # : 0 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:29.675 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@10 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@11 -- # hostsock=/var/tmp/host.sock 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@12 -- # loops=5 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # uuidgen 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # ns1uuid=6679aa19-4a7c-44f1-ad1d-fe8ba566f844 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # uuidgen 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # ns2uuid=42d8995f-2473-4ffe-b307-aa2189028864 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@16 -- # SUBSYSNQN=nqn.2016-06.io.spdk:cnode1 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@17 -- # HOSTNQN1=nqn.2016-06.io.spdk:host1 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@18 -- # HOSTNQN2=nqn.2016-06.io.spdk:host2 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # uuidgen 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # HOSTID=a28a47c0-6fb7-4291-9e7c-aa044e33cc41 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@50 -- # nvmftestinit 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@476 -- # prepare_net_devs 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@438 -- # local -g is_hw=no 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # remove_spdk_ns 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@309 -- # xtrace_disable 00:14:29.675 19:10:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # pci_devs=() 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # net_devs=() 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # e810=() 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # local -ga e810 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # x722=() 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # local -ga x722 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # mlx=() 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # local -ga mlx 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:14:37.808 Found 0000:31:00.0 (0x8086 - 0x159b) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:14:37.808 Found 0000:31:00.1 (0x8086 - 0x159b) 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:37.808 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:14:37.809 Found net devices under 0000:31:00.0: cvl_0_0 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:14:37.809 Found net devices under 0000:31:00.1: cvl_0_1 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # is_hw=yes 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:37.809 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:38.070 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:38.070 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.565 ms 00:14:38.070 00:14:38.070 --- 10.0.0.2 ping statistics --- 00:14:38.070 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:38.070 rtt min/avg/max/mdev = 0.565/0.565/0.565/0.000 ms 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:38.070 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:38.070 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.321 ms 00:14:38.070 00:14:38.070 --- 10.0.0.1 ping statistics --- 00:14:38.070 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:38.070 rtt min/avg/max/mdev = 0.321/0.321/0.321/0.000 ms 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@450 -- # return 0 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@51 -- # nvmfappstart 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@509 -- # nvmfpid=3795251 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@510 -- # waitforlisten 3795251 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # '[' -z 3795251 ']' 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:38.070 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:38.070 19:10:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:14:38.070 [2024-11-26 19:10:50.594743] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:14:38.070 [2024-11-26 19:10:50.594810] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:38.070 [2024-11-26 19:10:50.683917] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:38.498 [2024-11-26 19:10:50.724786] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:38.498 [2024-11-26 19:10:50.724824] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:38.498 [2024-11-26 19:10:50.724834] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:38.498 [2024-11-26 19:10:50.724842] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:38.498 [2024-11-26 19:10:50.724849] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:38.498 [2024-11-26 19:10:50.725532] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:38.778 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:38.778 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@868 -- # return 0 00:14:38.778 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:14:38.778 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:38.778 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:14:39.039 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:39.039 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:14:39.039 [2024-11-26 19:10:51.584574] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:39.039 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@55 -- # MALLOC_BDEV_SIZE=64 00:14:39.039 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@56 -- # MALLOC_BLOCK_SIZE=512 00:14:39.039 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:14:39.298 Malloc1 00:14:39.298 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:14:39.559 Malloc2 00:14:39.559 19:10:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:14:39.559 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 00:14:39.819 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:39.819 [2024-11-26 19:10:52.434623] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:40.079 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@67 -- # connect 00:14:40.079 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I a28a47c0-6fb7-4291-9e7c-aa044e33cc41 -a 10.0.0.2 -s 4420 -i 4 00:14:40.079 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 00:14:40.079 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1202 -- # local i=0 00:14:40.079 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:14:40.079 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:14:40.079 19:10:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1209 -- # sleep 2 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # return 0 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@68 -- # ns_is_visible 0x1 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:42.620 [ 0]:0x1 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ed69f3147cce42cf9da915f42445280b 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ed69f3147cce42cf9da915f42445280b != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@72 -- # ns_is_visible 0x1 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:42.620 [ 0]:0x1 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:42.620 19:10:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ed69f3147cce42cf9da915f42445280b 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ed69f3147cce42cf9da915f42445280b != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@73 -- # ns_is_visible 0x2 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:14:42.620 [ 1]:0x2 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=2037cd2c3a7d4655a2e1865dc9cdcae7 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 2037cd2c3a7d4655a2e1865dc9cdcae7 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@75 -- # disconnect 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:14:42.620 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:14:42.620 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:14:42.896 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 --no-auto-visible 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@83 -- # connect 1 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I a28a47c0-6fb7-4291-9e7c-aa044e33cc41 -a 10.0.0.2 -s 4420 -i 4 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 1 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1202 -- # local i=0 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1204 -- # [[ -n 1 ]] 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # nvme_device_counter=1 00:14:43.157 19:10:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1209 -- # sleep 2 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # return 0 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@84 -- # NOT ns_is_visible 0x1 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@85 -- # ns_is_visible 0x2 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:14:45.704 [ 0]:0x2 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=2037cd2c3a7d4655a2e1865dc9cdcae7 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 2037cd2c3a7d4655a2e1865dc9cdcae7 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:45.704 19:10:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@89 -- # ns_is_visible 0x1 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:45.704 [ 0]:0x1 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ed69f3147cce42cf9da915f42445280b 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ed69f3147cce42cf9da915f42445280b != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@90 -- # ns_is_visible 0x2 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:14:45.704 [ 1]:0x2 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=2037cd2c3a7d4655a2e1865dc9cdcae7 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 2037cd2c3a7d4655a2e1865dc9cdcae7 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:45.704 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@94 -- # NOT ns_is_visible 0x1 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@95 -- # ns_is_visible 0x2 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:14:45.964 [ 0]:0x2 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=2037cd2c3a7d4655a2e1865dc9cdcae7 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 2037cd2c3a7d4655a2e1865dc9cdcae7 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:45.964 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@97 -- # disconnect 00:14:45.965 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:14:45.965 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:14:45.965 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:14:46.224 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@101 -- # connect 2 00:14:46.224 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I a28a47c0-6fb7-4291-9e7c-aa044e33cc41 -a 10.0.0.2 -s 4420 -i 4 00:14:46.485 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 2 00:14:46.485 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1202 -- # local i=0 00:14:46.485 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:14:46.485 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1204 -- # [[ -n 2 ]] 00:14:46.485 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # nvme_device_counter=2 00:14:46.485 19:10:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1209 -- # sleep 2 00:14:48.401 19:11:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:14:48.401 19:11:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:14:48.401 19:11:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:14:48.401 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # nvme_devices=2 00:14:48.401 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:14:48.401 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # return 0 00:14:48.401 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:14:48.401 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@102 -- # ns_is_visible 0x1 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:48.660 [ 0]:0x1 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:48.660 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ed69f3147cce42cf9da915f42445280b 00:14:48.661 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ed69f3147cce42cf9da915f42445280b != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:48.661 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@103 -- # ns_is_visible 0x2 00:14:48.661 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:48.661 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:14:48.921 [ 1]:0x2 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=2037cd2c3a7d4655a2e1865dc9cdcae7 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 2037cd2c3a7d4655a2e1865dc9cdcae7 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@106 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@107 -- # NOT ns_is_visible 0x1 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:48.921 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@108 -- # ns_is_visible 0x2 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:14:49.181 [ 0]:0x2 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=2037cd2c3a7d4655a2e1865dc9cdcae7 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 2037cd2c3a7d4655a2e1865dc9cdcae7 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@111 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:49.181 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:14:49.182 [2024-11-26 19:11:01.773286] nvmf_rpc.c:1873:nvmf_rpc_ns_visible_paused: *ERROR*: Unable to add/remove nqn.2016-06.io.spdk:host1 to namespace ID 2 00:14:49.182 request: 00:14:49.182 { 00:14:49.182 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:14:49.182 "nsid": 2, 00:14:49.182 "host": "nqn.2016-06.io.spdk:host1", 00:14:49.182 "method": "nvmf_ns_remove_host", 00:14:49.182 "req_id": 1 00:14:49.182 } 00:14:49.182 Got JSON-RPC error response 00:14:49.182 response: 00:14:49.182 { 00:14:49.182 "code": -32602, 00:14:49.182 "message": "Invalid parameters" 00:14:49.182 } 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@112 -- # NOT ns_is_visible 0x1 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:49.182 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@113 -- # ns_is_visible 0x2 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:14:49.442 [ 0]:0x2 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=2037cd2c3a7d4655a2e1865dc9cdcae7 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 2037cd2c3a7d4655a2e1865dc9cdcae7 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@114 -- # disconnect 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:14:49.442 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@118 -- # hostpid=3797685 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@119 -- # trap 'killprocess $hostpid; nvmftestfini' SIGINT SIGTERM EXIT 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@117 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -r /var/tmp/host.sock -m 2 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@121 -- # waitforlisten 3797685 /var/tmp/host.sock 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # '[' -z 3797685 ']' 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/host.sock 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:14:49.442 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:49.442 19:11:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:14:49.442 [2024-11-26 19:11:02.032809] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:14:49.442 [2024-11-26 19:11:02.032870] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3797685 ] 00:14:49.701 [2024-11-26 19:11:02.126957] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:49.701 [2024-11-26 19:11:02.164225] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:50.270 19:11:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:50.270 19:11:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@868 -- # return 0 00:14:50.270 19:11:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@122 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:14:50.531 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:14:50.805 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # uuid2nguid 6679aa19-4a7c-44f1-ad1d-fe8ba566f844 00:14:50.805 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:14:50.805 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g 6679AA194A7C44F1AD1DFE8BA566F844 -i 00:14:50.805 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # uuid2nguid 42d8995f-2473-4ffe-b307-aa2189028864 00:14:50.805 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:14:50.805 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 -g 42D8995F24734FFEB307AA2189028864 -i 00:14:51.066 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:14:51.066 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@127 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host2 00:14:51.325 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@129 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:14:51.326 19:11:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:14:51.895 nvme0n1 00:14:51.895 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@131 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:14:51.895 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:14:52.155 nvme1n2 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # hostrpc bdev_get_bdevs 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # jq -r '.[].name' 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # sort 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # xargs 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # [[ nvme0n1 nvme1n2 == \n\v\m\e\0\n\1\ \n\v\m\e\1\n\2 ]] 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # hostrpc bdev_get_bdevs -b nvme0n1 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # jq -r '.[].uuid' 00:14:52.155 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme0n1 00:14:52.414 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # [[ 6679aa19-4a7c-44f1-ad1d-fe8ba566f844 == \6\6\7\9\a\a\1\9\-\4\a\7\c\-\4\4\f\1\-\a\d\1\d\-\f\e\8\b\a\5\6\6\f\8\4\4 ]] 00:14:52.414 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # hostrpc bdev_get_bdevs -b nvme1n2 00:14:52.415 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # jq -r '.[].uuid' 00:14:52.415 19:11:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme1n2 00:14:52.674 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # [[ 42d8995f-2473-4ffe-b307-aa2189028864 == \4\2\d\8\9\9\5\f\-\2\4\7\3\-\4\f\f\e\-\b\3\0\7\-\a\a\2\1\8\9\0\2\8\8\6\4 ]] 00:14:52.674 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@137 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:14:52.674 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@138 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@141 -- # uuid2nguid 6679aa19-4a7c-44f1-ad1d-fe8ba566f844 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@141 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 6679AA194A7C44F1AD1DFE8BA566F844 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 6679AA194A7C44F1AD1DFE8BA566F844 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:14:52.934 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 6679AA194A7C44F1AD1DFE8BA566F844 00:14:52.934 [2024-11-26 19:11:05.539645] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: invalid 00:14:52.934 [2024-11-26 19:11:05.539678] subsystem.c:2156:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode1: bdev invalid cannot be opened, error=-19 00:14:52.934 [2024-11-26 19:11:05.539688] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:52.934 request: 00:14:52.934 { 00:14:52.934 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:14:52.934 "namespace": { 00:14:52.934 "bdev_name": "invalid", 00:14:52.934 "nsid": 1, 00:14:52.934 "nguid": "6679AA194A7C44F1AD1DFE8BA566F844", 00:14:52.934 "no_auto_visible": false, 00:14:52.934 "hide_metadata": false 00:14:52.934 }, 00:14:52.934 "method": "nvmf_subsystem_add_ns", 00:14:52.934 "req_id": 1 00:14:52.934 } 00:14:52.934 Got JSON-RPC error response 00:14:52.934 response: 00:14:52.934 { 00:14:52.934 "code": -32602, 00:14:52.934 "message": "Invalid parameters" 00:14:52.935 } 00:14:52.935 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:14:52.935 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:52.935 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:52.935 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:53.194 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@142 -- # uuid2nguid 6679aa19-4a7c-44f1-ad1d-fe8ba566f844 00:14:53.194 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:14:53.194 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@142 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g 6679AA194A7C44F1AD1DFE8BA566F844 -i 00:14:53.194 19:11:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@143 -- # sleep 2s 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # hostrpc bdev_get_bdevs 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # jq length 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # (( 0 == 0 )) 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@146 -- # killprocess 3797685 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # '[' -z 3797685 ']' 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@958 -- # kill -0 3797685 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # uname 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3797685 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3797685' 00:14:55.734 killing process with pid 3797685 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@973 -- # kill 3797685 00:14:55.734 19:11:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@978 -- # wait 3797685 00:14:55.734 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:14:55.994 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@150 -- # nvmftestfini 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@516 -- # nvmfcleanup 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@121 -- # sync 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@124 -- # set +e 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:55.995 rmmod nvme_tcp 00:14:55.995 rmmod nvme_fabrics 00:14:55.995 rmmod nvme_keyring 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@128 -- # set -e 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@129 -- # return 0 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@517 -- # '[' -n 3795251 ']' 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@518 -- # killprocess 3795251 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # '[' -z 3795251 ']' 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@958 -- # kill -0 3795251 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # uname 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3795251 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3795251' 00:14:55.995 killing process with pid 3795251 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@973 -- # kill 3795251 00:14:55.995 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@978 -- # wait 3795251 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@297 -- # iptr 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@791 -- # iptables-save 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@791 -- # iptables-restore 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:56.255 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:56.256 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:56.256 19:11:08 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:14:58.165 00:14:58.165 real 0m28.961s 00:14:58.165 user 0m31.782s 00:14:58.165 sys 0m8.823s 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:14:58.165 ************************************ 00:14:58.165 END TEST nvmf_ns_masking 00:14:58.165 ************************************ 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@27 -- # [[ 1 -eq 1 ]] 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@28 -- # run_test nvmf_nvme_cli /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:58.165 19:11:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:14:58.426 ************************************ 00:14:58.426 START TEST nvmf_nvme_cli 00:14:58.426 ************************************ 00:14:58.426 19:11:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:14:58.426 * Looking for test storage... 00:14:58.426 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:58.426 19:11:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:14:58.426 19:11:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1693 -- # lcov --version 00:14:58.426 19:11:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # IFS=.-: 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # read -ra ver1 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # IFS=.-: 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # read -ra ver2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@338 -- # local 'op=<' 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@340 -- # ver1_l=2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@341 -- # ver2_l=1 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@344 -- # case "$op" in 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@345 -- # : 1 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # decimal 1 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=1 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 1 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # ver1[v]=1 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # decimal 2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # ver2[v]=2 00:14:58.426 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # return 0 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:14:58.427 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:58.427 --rc genhtml_branch_coverage=1 00:14:58.427 --rc genhtml_function_coverage=1 00:14:58.427 --rc genhtml_legend=1 00:14:58.427 --rc geninfo_all_blocks=1 00:14:58.427 --rc geninfo_unexecuted_blocks=1 00:14:58.427 00:14:58.427 ' 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:14:58.427 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:58.427 --rc genhtml_branch_coverage=1 00:14:58.427 --rc genhtml_function_coverage=1 00:14:58.427 --rc genhtml_legend=1 00:14:58.427 --rc geninfo_all_blocks=1 00:14:58.427 --rc geninfo_unexecuted_blocks=1 00:14:58.427 00:14:58.427 ' 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:14:58.427 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:58.427 --rc genhtml_branch_coverage=1 00:14:58.427 --rc genhtml_function_coverage=1 00:14:58.427 --rc genhtml_legend=1 00:14:58.427 --rc geninfo_all_blocks=1 00:14:58.427 --rc geninfo_unexecuted_blocks=1 00:14:58.427 00:14:58.427 ' 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:14:58.427 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:58.427 --rc genhtml_branch_coverage=1 00:14:58.427 --rc genhtml_function_coverage=1 00:14:58.427 --rc genhtml_legend=1 00:14:58.427 --rc geninfo_all_blocks=1 00:14:58.427 --rc geninfo_unexecuted_blocks=1 00:14:58.427 00:14:58.427 ' 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # uname -s 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:58.427 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@15 -- # shopt -s extglob 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@5 -- # export PATH 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@51 -- # : 0 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:58.688 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@11 -- # MALLOC_BDEV_SIZE=64 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@14 -- # devs=() 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@16 -- # nvmftestinit 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@476 -- # prepare_net_devs 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@438 -- # local -g is_hw=no 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # remove_spdk_ns 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@309 -- # xtrace_disable 00:14:58.688 19:11:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # pci_devs=() 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # net_devs=() 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # e810=() 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # local -ga e810 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # x722=() 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # local -ga x722 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # mlx=() 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # local -ga mlx 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:15:06.824 Found 0000:31:00.0 (0x8086 - 0x159b) 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:06.824 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:15:06.825 Found 0000:31:00.1 (0x8086 - 0x159b) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:15:06.825 Found net devices under 0000:31:00.0: cvl_0_0 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:15:06.825 Found net devices under 0000:31:00.1: cvl_0_1 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # is_hw=yes 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:06.825 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:07.085 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:07.085 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.699 ms 00:15:07.085 00:15:07.085 --- 10.0.0.2 ping statistics --- 00:15:07.085 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:07.085 rtt min/avg/max/mdev = 0.699/0.699/0.699/0.000 ms 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:07.085 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:07.085 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.278 ms 00:15:07.085 00:15:07.085 --- 10.0.0.1 ping statistics --- 00:15:07.085 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:07.085 rtt min/avg/max/mdev = 0.278/0.278/0.278/0.000 ms 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@450 -- # return 0 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@17 -- # nvmfappstart -m 0xF 00:15:07.085 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@726 -- # xtrace_disable 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@509 -- # nvmfpid=3803743 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@510 -- # waitforlisten 3803743 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@835 -- # '[' -z 3803743 ']' 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:07.086 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:07.086 19:11:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.086 [2024-11-26 19:11:19.591631] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:15:07.086 [2024-11-26 19:11:19.591729] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:07.086 [2024-11-26 19:11:19.685124] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:07.346 [2024-11-26 19:11:19.727724] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:07.346 [2024-11-26 19:11:19.727761] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:07.346 [2024-11-26 19:11:19.727769] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:07.346 [2024-11-26 19:11:19.727776] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:07.346 [2024-11-26 19:11:19.727782] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:07.346 [2024-11-26 19:11:19.729412] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:07.346 [2024-11-26 19:11:19.729531] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:07.347 [2024-11-26 19:11:19.729692] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:07.347 [2024-11-26 19:11:19.729693] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@868 -- # return 0 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@732 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 [2024-11-26 19:11:20.449369] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@21 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 Malloc0 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 Malloc1 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME -d SPDK_Controller1 -i 291 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:07.941 [2024-11-26 19:11:20.549488] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@28 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.941 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:08.201 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.201 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@30 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -a 10.0.0.2 -s 4420 00:15:08.201 00:15:08.201 Discovery Log Number of Records 2, Generation counter 2 00:15:08.201 =====Discovery Log Entry 0====== 00:15:08.201 trtype: tcp 00:15:08.201 adrfam: ipv4 00:15:08.201 subtype: current discovery subsystem 00:15:08.201 treq: not required 00:15:08.201 portid: 0 00:15:08.201 trsvcid: 4420 00:15:08.201 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:15:08.201 traddr: 10.0.0.2 00:15:08.201 eflags: explicit discovery connections, duplicate discovery information 00:15:08.201 sectype: none 00:15:08.201 =====Discovery Log Entry 1====== 00:15:08.201 trtype: tcp 00:15:08.201 adrfam: ipv4 00:15:08.201 subtype: nvme subsystem 00:15:08.201 treq: not required 00:15:08.201 portid: 0 00:15:08.201 trsvcid: 4420 00:15:08.201 subnqn: nqn.2016-06.io.spdk:cnode1 00:15:08.201 traddr: 10.0.0.2 00:15:08.201 eflags: none 00:15:08.201 sectype: none 00:15:08.201 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # devs=($(get_nvme_devs)) 00:15:08.201 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # get_nvme_devs 00:15:08.201 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # local dev _ 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@549 -- # nvme list 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ Node == /dev/nvme* ]] 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ --------------------- == /dev/nvme* ]] 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # nvme_num_before_connection=0 00:15:08.202 19:11:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@32 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:15:10.114 19:11:22 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@34 -- # waitforserial SPDKISFASTANDAWESOME 2 00:15:10.114 19:11:22 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1202 -- # local i=0 00:15:10.114 19:11:22 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:15:10.114 19:11:22 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1204 -- # [[ -n 2 ]] 00:15:10.114 19:11:22 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1205 -- # nvme_device_counter=2 00:15:10.114 19:11:22 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1209 -- # sleep 2 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1211 -- # nvme_devices=2 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1212 -- # return 0 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # get_nvme_devs 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # local dev _ 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@549 -- # nvme list 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ Node == /dev/nvme* ]] 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ --------------------- == /dev/nvme* ]] 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n1 00:15:12.025 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n2 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # [[ -z /dev/nvme0n1 00:15:12.026 /dev/nvme0n2 ]] 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # devs=($(get_nvme_devs)) 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # get_nvme_devs 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # local dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@549 -- # nvme list 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ Node == /dev/nvme* ]] 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ --------------------- == /dev/nvme* ]] 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n1 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n2 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # nvme_num=2 00:15:12.026 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@60 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:15:12.286 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@61 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1223 -- # local i=0 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1235 -- # return 0 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@62 -- # (( nvme_num <= nvme_num_before_connection )) 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@67 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:15:12.286 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@70 -- # nvmftestfini 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@516 -- # nvmfcleanup 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@121 -- # sync 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@124 -- # set +e 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@125 -- # for i in {1..20} 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:15:12.546 rmmod nvme_tcp 00:15:12.546 rmmod nvme_fabrics 00:15:12.546 rmmod nvme_keyring 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@128 -- # set -e 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@129 -- # return 0 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@517 -- # '[' -n 3803743 ']' 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@518 -- # killprocess 3803743 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@954 -- # '[' -z 3803743 ']' 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@958 -- # kill -0 3803743 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@959 -- # uname 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:12.546 19:11:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3803743 00:15:12.546 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:12.546 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:12.546 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3803743' 00:15:12.546 killing process with pid 3803743 00:15:12.546 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@973 -- # kill 3803743 00:15:12.546 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@978 -- # wait 3803743 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@297 -- # iptr 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@791 -- # iptables-save 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@791 -- # iptables-restore 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@302 -- # remove_spdk_ns 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:12.807 19:11:25 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:15:14.719 00:15:14.719 real 0m16.455s 00:15:14.719 user 0m24.360s 00:15:14.719 sys 0m7.016s 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:15:14.719 ************************************ 00:15:14.719 END TEST nvmf_nvme_cli 00:15:14.719 ************************************ 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@30 -- # [[ 1 -eq 1 ]] 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@31 -- # run_test nvmf_vfio_user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:14.719 19:11:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:14.980 ************************************ 00:15:14.980 START TEST nvmf_vfio_user 00:15:14.980 ************************************ 00:15:14.980 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:15:14.980 * Looking for test storage... 00:15:14.980 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:14.980 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:15:14.980 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:15:14.980 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1693 -- # lcov --version 00:15:14.980 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:15:14.980 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # IFS=.-: 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # read -ra ver1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # IFS=.-: 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # read -ra ver2 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@338 -- # local 'op=<' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@340 -- # ver1_l=2 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@341 -- # ver2_l=1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@344 -- # case "$op" in 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@345 -- # : 1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # decimal 1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # ver1[v]=1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # decimal 2 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=2 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 2 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # ver2[v]=2 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # return 0 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:15:14.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.981 --rc genhtml_branch_coverage=1 00:15:14.981 --rc genhtml_function_coverage=1 00:15:14.981 --rc genhtml_legend=1 00:15:14.981 --rc geninfo_all_blocks=1 00:15:14.981 --rc geninfo_unexecuted_blocks=1 00:15:14.981 00:15:14.981 ' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:15:14.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.981 --rc genhtml_branch_coverage=1 00:15:14.981 --rc genhtml_function_coverage=1 00:15:14.981 --rc genhtml_legend=1 00:15:14.981 --rc geninfo_all_blocks=1 00:15:14.981 --rc geninfo_unexecuted_blocks=1 00:15:14.981 00:15:14.981 ' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:15:14.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.981 --rc genhtml_branch_coverage=1 00:15:14.981 --rc genhtml_function_coverage=1 00:15:14.981 --rc genhtml_legend=1 00:15:14.981 --rc geninfo_all_blocks=1 00:15:14.981 --rc geninfo_unexecuted_blocks=1 00:15:14.981 00:15:14.981 ' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:15:14.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.981 --rc genhtml_branch_coverage=1 00:15:14.981 --rc genhtml_function_coverage=1 00:15:14.981 --rc genhtml_legend=1 00:15:14.981 --rc geninfo_all_blocks=1 00:15:14.981 --rc geninfo_unexecuted_blocks=1 00:15:14.981 00:15:14.981 ' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # uname -s 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@15 -- # shopt -s extglob 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@5 -- # export PATH 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@51 -- # : 0 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:14.981 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:14.981 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:14.982 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:14.982 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@12 -- # MALLOC_BDEV_SIZE=64 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@14 -- # NUM_DEVICES=2 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@47 -- # rm -rf /var/run/vfio-user 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@103 -- # setup_nvmf_vfio_user '' '' 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args= 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local transport_args= 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=3805323 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 3805323' 00:15:15.242 Process pid: 3805323 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 3805323 00:15:15.242 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # '[' -z 3805323 ']' 00:15:15.243 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' 00:15:15.243 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:15.243 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:15.243 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:15.243 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:15.243 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:15.243 19:11:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:15:15.243 [2024-11-26 19:11:27.673310] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:15:15.243 [2024-11-26 19:11:27.673361] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:15.243 [2024-11-26 19:11:27.755386] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:15.243 [2024-11-26 19:11:27.791261] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:15.243 [2024-11-26 19:11:27.791295] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:15.243 [2024-11-26 19:11:27.791303] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:15.243 [2024-11-26 19:11:27.791310] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:15.243 [2024-11-26 19:11:27.791316] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:15.243 [2024-11-26 19:11:27.792857] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:15.243 [2024-11-26 19:11:27.792998] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:15.243 [2024-11-26 19:11:27.793235] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:15.243 [2024-11-26 19:11:27.793236] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:16.182 19:11:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:16.182 19:11:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@868 -- # return 0 00:15:16.182 19:11:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:15:17.121 19:11:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER 00:15:17.121 19:11:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:15:17.121 19:11:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:15:17.121 19:11:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:15:17.121 19:11:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:15:17.121 19:11:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:15:17.382 Malloc1 00:15:17.382 19:11:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:15:17.643 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:15:17.643 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:15:17.903 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:15:17.903 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:15:17.903 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:15:18.163 Malloc2 00:15:18.163 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:15:18.423 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:15:18.423 19:11:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:15:18.686 19:11:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@104 -- # run_nvmf_vfio_user 00:15:18.686 19:11:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # seq 1 2 00:15:18.686 19:11:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:15:18.686 19:11:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user1/1 00:15:18.686 19:11:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode1 00:15:18.686 19:11:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -L nvme -L nvme_vfio -L vfio_pci 00:15:18.686 [2024-11-26 19:11:31.173209] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:15:18.686 [2024-11-26 19:11:31.173244] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3806042 ] 00:15:18.686 [2024-11-26 19:11:31.226027] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user1/1 00:15:18.686 [2024-11-26 19:11:31.235142] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:15:18.686 [2024-11-26 19:11:31.235165] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7fc53c2d7000 00:15:18.686 [2024-11-26 19:11:31.236146] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.237145] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.238154] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.239158] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.240164] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.241174] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.242175] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.243183] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:18.687 [2024-11-26 19:11:31.244191] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:15:18.687 [2024-11-26 19:11:31.244201] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7fc53c2cc000 00:15:18.687 [2024-11-26 19:11:31.245528] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:15:18.687 [2024-11-26 19:11:31.266023] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user1/1/cntrl Setup Successfully 00:15:18.687 [2024-11-26 19:11:31.266065] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to connect adminq (no timeout) 00:15:18.687 [2024-11-26 19:11:31.268317] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:15:18.687 [2024-11-26 19:11:31.268371] nvme_pcie_common.c: 159:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:15:18.687 [2024-11-26 19:11:31.268463] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for connect adminq (no timeout) 00:15:18.687 [2024-11-26 19:11:31.268481] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read vs (no timeout) 00:15:18.687 [2024-11-26 19:11:31.268487] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read vs wait for vs (no timeout) 00:15:18.687 [2024-11-26 19:11:31.269313] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x8, value 0x10300 00:15:18.687 [2024-11-26 19:11:31.269326] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read cap (no timeout) 00:15:18.687 [2024-11-26 19:11:31.269333] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read cap wait for cap (no timeout) 00:15:18.687 [2024-11-26 19:11:31.270316] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:15:18.687 [2024-11-26 19:11:31.270326] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to check en (no timeout) 00:15:18.687 [2024-11-26 19:11:31.270334] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to check en wait for cc (timeout 15000 ms) 00:15:18.687 [2024-11-26 19:11:31.271323] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x0 00:15:18.687 [2024-11-26 19:11:31.271332] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:15:18.687 [2024-11-26 19:11:31.272327] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x0 00:15:18.687 [2024-11-26 19:11:31.272336] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CC.EN = 0 && CSTS.RDY = 0 00:15:18.687 [2024-11-26 19:11:31.272344] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to controller is disabled (timeout 15000 ms) 00:15:18.687 [2024-11-26 19:11:31.272352] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:15:18.687 [2024-11-26 19:11:31.272461] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Setting CC.EN = 1 00:15:18.687 [2024-11-26 19:11:31.272466] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:15:18.687 [2024-11-26 19:11:31.272472] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x28, value 0x2000003c0000 00:15:18.687 [2024-11-26 19:11:31.273334] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x30, value 0x2000003be000 00:15:18.687 [2024-11-26 19:11:31.274338] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x24, value 0xff00ff 00:15:18.687 [2024-11-26 19:11:31.275343] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:15:18.687 [2024-11-26 19:11:31.276343] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:18.687 [2024-11-26 19:11:31.276414] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:15:18.687 [2024-11-26 19:11:31.277352] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x1 00:15:18.687 [2024-11-26 19:11:31.277361] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:15:18.687 [2024-11-26 19:11:31.277366] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to reset admin queue (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277388] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify controller (no timeout) 00:15:18.687 [2024-11-26 19:11:31.277396] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify controller (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277416] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:15:18.687 [2024-11-26 19:11:31.277421] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:18.687 [2024-11-26 19:11:31.277425] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:18.687 [2024-11-26 19:11:31.277439] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:18.687 [2024-11-26 19:11:31.277480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:15:18.687 [2024-11-26 19:11:31.277490] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] transport max_xfer_size 131072 00:15:18.687 [2024-11-26 19:11:31.277495] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] MDTS max_xfer_size 131072 00:15:18.687 [2024-11-26 19:11:31.277500] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CNTLID 0x0001 00:15:18.687 [2024-11-26 19:11:31.277505] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:15:18.687 [2024-11-26 19:11:31.277509] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] transport max_sges 1 00:15:18.687 [2024-11-26 19:11:31.277516] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] fuses compare and write: 1 00:15:18.687 [2024-11-26 19:11:31.277521] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to configure AER (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277530] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for configure aer (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277540] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:15:18.687 [2024-11-26 19:11:31.277554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:15:18.687 [2024-11-26 19:11:31.277565] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:15:18.687 [2024-11-26 19:11:31.277574] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:15:18.687 [2024-11-26 19:11:31.277583] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:15:18.687 [2024-11-26 19:11:31.277591] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:15:18.687 [2024-11-26 19:11:31.277596] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set keep alive timeout (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277606] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277615] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:15:18.687 [2024-11-26 19:11:31.277622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:15:18.687 [2024-11-26 19:11:31.277628] nvme_ctrlr.c:3047:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Controller adjusted keep alive timeout to 0 ms 00:15:18.687 [2024-11-26 19:11:31.277633] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify controller iocs specific (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277642] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set number of queues (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277649] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for set number of queues (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277658] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:15:18.687 [2024-11-26 19:11:31.277668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:15:18.687 [2024-11-26 19:11:31.277730] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify active ns (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277739] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify active ns (timeout 30000 ms) 00:15:18.687 [2024-11-26 19:11:31.277747] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:15:18.687 [2024-11-26 19:11:31.277751] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:15:18.687 [2024-11-26 19:11:31.277755] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:18.687 [2024-11-26 19:11:31.277761] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:15:18.687 [2024-11-26 19:11:31.277777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:15:18.687 [2024-11-26 19:11:31.277790] nvme_ctrlr.c:4735:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Namespace 1 was added 00:15:18.688 [2024-11-26 19:11:31.277802] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify ns (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277810] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify ns (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277817] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:15:18.688 [2024-11-26 19:11:31.277821] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:18.688 [2024-11-26 19:11:31.277825] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:18.688 [2024-11-26 19:11:31.277831] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.277854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.277878] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify namespace id descriptors (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277887] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277894] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:15:18.688 [2024-11-26 19:11:31.277898] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:18.688 [2024-11-26 19:11:31.277902] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:18.688 [2024-11-26 19:11:31.277908] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.277922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.277933] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify ns iocs specific (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277940] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set supported log pages (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277948] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set supported features (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277954] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set host behavior support feature (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277959] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set doorbell buffer config (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277965] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set host ID (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277970] nvme_ctrlr.c:3147:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] NVMe-oF transport - not sending Set Features - Host ID 00:15:18.688 [2024-11-26 19:11:31.277974] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to transport ready (timeout 30000 ms) 00:15:18.688 [2024-11-26 19:11:31.277980] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to ready (no timeout) 00:15:18.688 [2024-11-26 19:11:31.278000] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.278011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.278023] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.278030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.278041] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.278049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.278060] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.278070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.278083] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:15:18.688 [2024-11-26 19:11:31.278088] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:15:18.688 [2024-11-26 19:11:31.278092] nvme_pcie_common.c:1275:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:15:18.688 [2024-11-26 19:11:31.278096] nvme_pcie_common.c:1291:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:15:18.688 [2024-11-26 19:11:31.278099] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:15:18.688 [2024-11-26 19:11:31.278105] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:15:18.688 [2024-11-26 19:11:31.278113] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:15:18.688 [2024-11-26 19:11:31.278118] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:15:18.688 [2024-11-26 19:11:31.278121] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:18.688 [2024-11-26 19:11:31.278127] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.278135] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:15:18.688 [2024-11-26 19:11:31.278139] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:18.688 [2024-11-26 19:11:31.278142] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:18.688 [2024-11-26 19:11:31.278148] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.278156] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:15:18.688 [2024-11-26 19:11:31.278161] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:15:18.688 [2024-11-26 19:11:31.278164] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:18.688 [2024-11-26 19:11:31.278170] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:15:18.688 [2024-11-26 19:11:31.278178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.278191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.278203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:15:18.688 [2024-11-26 19:11:31.278211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:15:18.688 ===================================================== 00:15:18.688 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:15:18.688 ===================================================== 00:15:18.688 Controller Capabilities/Features 00:15:18.688 ================================ 00:15:18.688 Vendor ID: 4e58 00:15:18.688 Subsystem Vendor ID: 4e58 00:15:18.688 Serial Number: SPDK1 00:15:18.688 Model Number: SPDK bdev Controller 00:15:18.688 Firmware Version: 25.01 00:15:18.688 Recommended Arb Burst: 6 00:15:18.688 IEEE OUI Identifier: 8d 6b 50 00:15:18.688 Multi-path I/O 00:15:18.688 May have multiple subsystem ports: Yes 00:15:18.688 May have multiple controllers: Yes 00:15:18.688 Associated with SR-IOV VF: No 00:15:18.688 Max Data Transfer Size: 131072 00:15:18.688 Max Number of Namespaces: 32 00:15:18.688 Max Number of I/O Queues: 127 00:15:18.688 NVMe Specification Version (VS): 1.3 00:15:18.688 NVMe Specification Version (Identify): 1.3 00:15:18.688 Maximum Queue Entries: 256 00:15:18.688 Contiguous Queues Required: Yes 00:15:18.688 Arbitration Mechanisms Supported 00:15:18.688 Weighted Round Robin: Not Supported 00:15:18.688 Vendor Specific: Not Supported 00:15:18.688 Reset Timeout: 15000 ms 00:15:18.688 Doorbell Stride: 4 bytes 00:15:18.688 NVM Subsystem Reset: Not Supported 00:15:18.688 Command Sets Supported 00:15:18.688 NVM Command Set: Supported 00:15:18.688 Boot Partition: Not Supported 00:15:18.688 Memory Page Size Minimum: 4096 bytes 00:15:18.688 Memory Page Size Maximum: 4096 bytes 00:15:18.688 Persistent Memory Region: Not Supported 00:15:18.688 Optional Asynchronous Events Supported 00:15:18.688 Namespace Attribute Notices: Supported 00:15:18.688 Firmware Activation Notices: Not Supported 00:15:18.688 ANA Change Notices: Not Supported 00:15:18.688 PLE Aggregate Log Change Notices: Not Supported 00:15:18.688 LBA Status Info Alert Notices: Not Supported 00:15:18.688 EGE Aggregate Log Change Notices: Not Supported 00:15:18.688 Normal NVM Subsystem Shutdown event: Not Supported 00:15:18.688 Zone Descriptor Change Notices: Not Supported 00:15:18.688 Discovery Log Change Notices: Not Supported 00:15:18.688 Controller Attributes 00:15:18.688 128-bit Host Identifier: Supported 00:15:18.688 Non-Operational Permissive Mode: Not Supported 00:15:18.688 NVM Sets: Not Supported 00:15:18.688 Read Recovery Levels: Not Supported 00:15:18.688 Endurance Groups: Not Supported 00:15:18.688 Predictable Latency Mode: Not Supported 00:15:18.688 Traffic Based Keep ALive: Not Supported 00:15:18.688 Namespace Granularity: Not Supported 00:15:18.688 SQ Associations: Not Supported 00:15:18.688 UUID List: Not Supported 00:15:18.688 Multi-Domain Subsystem: Not Supported 00:15:18.688 Fixed Capacity Management: Not Supported 00:15:18.688 Variable Capacity Management: Not Supported 00:15:18.688 Delete Endurance Group: Not Supported 00:15:18.688 Delete NVM Set: Not Supported 00:15:18.688 Extended LBA Formats Supported: Not Supported 00:15:18.688 Flexible Data Placement Supported: Not Supported 00:15:18.688 00:15:18.688 Controller Memory Buffer Support 00:15:18.688 ================================ 00:15:18.688 Supported: No 00:15:18.688 00:15:18.688 Persistent Memory Region Support 00:15:18.688 ================================ 00:15:18.688 Supported: No 00:15:18.688 00:15:18.688 Admin Command Set Attributes 00:15:18.689 ============================ 00:15:18.689 Security Send/Receive: Not Supported 00:15:18.689 Format NVM: Not Supported 00:15:18.689 Firmware Activate/Download: Not Supported 00:15:18.689 Namespace Management: Not Supported 00:15:18.689 Device Self-Test: Not Supported 00:15:18.689 Directives: Not Supported 00:15:18.689 NVMe-MI: Not Supported 00:15:18.689 Virtualization Management: Not Supported 00:15:18.689 Doorbell Buffer Config: Not Supported 00:15:18.689 Get LBA Status Capability: Not Supported 00:15:18.689 Command & Feature Lockdown Capability: Not Supported 00:15:18.689 Abort Command Limit: 4 00:15:18.689 Async Event Request Limit: 4 00:15:18.689 Number of Firmware Slots: N/A 00:15:18.689 Firmware Slot 1 Read-Only: N/A 00:15:18.689 Firmware Activation Without Reset: N/A 00:15:18.689 Multiple Update Detection Support: N/A 00:15:18.689 Firmware Update Granularity: No Information Provided 00:15:18.689 Per-Namespace SMART Log: No 00:15:18.689 Asymmetric Namespace Access Log Page: Not Supported 00:15:18.689 Subsystem NQN: nqn.2019-07.io.spdk:cnode1 00:15:18.689 Command Effects Log Page: Supported 00:15:18.689 Get Log Page Extended Data: Supported 00:15:18.689 Telemetry Log Pages: Not Supported 00:15:18.689 Persistent Event Log Pages: Not Supported 00:15:18.689 Supported Log Pages Log Page: May Support 00:15:18.689 Commands Supported & Effects Log Page: Not Supported 00:15:18.689 Feature Identifiers & Effects Log Page:May Support 00:15:18.689 NVMe-MI Commands & Effects Log Page: May Support 00:15:18.689 Data Area 4 for Telemetry Log: Not Supported 00:15:18.689 Error Log Page Entries Supported: 128 00:15:18.689 Keep Alive: Supported 00:15:18.689 Keep Alive Granularity: 10000 ms 00:15:18.689 00:15:18.689 NVM Command Set Attributes 00:15:18.689 ========================== 00:15:18.689 Submission Queue Entry Size 00:15:18.689 Max: 64 00:15:18.689 Min: 64 00:15:18.689 Completion Queue Entry Size 00:15:18.689 Max: 16 00:15:18.689 Min: 16 00:15:18.689 Number of Namespaces: 32 00:15:18.689 Compare Command: Supported 00:15:18.689 Write Uncorrectable Command: Not Supported 00:15:18.689 Dataset Management Command: Supported 00:15:18.689 Write Zeroes Command: Supported 00:15:18.689 Set Features Save Field: Not Supported 00:15:18.689 Reservations: Not Supported 00:15:18.689 Timestamp: Not Supported 00:15:18.689 Copy: Supported 00:15:18.689 Volatile Write Cache: Present 00:15:18.689 Atomic Write Unit (Normal): 1 00:15:18.689 Atomic Write Unit (PFail): 1 00:15:18.689 Atomic Compare & Write Unit: 1 00:15:18.689 Fused Compare & Write: Supported 00:15:18.689 Scatter-Gather List 00:15:18.689 SGL Command Set: Supported (Dword aligned) 00:15:18.689 SGL Keyed: Not Supported 00:15:18.689 SGL Bit Bucket Descriptor: Not Supported 00:15:18.689 SGL Metadata Pointer: Not Supported 00:15:18.689 Oversized SGL: Not Supported 00:15:18.689 SGL Metadata Address: Not Supported 00:15:18.689 SGL Offset: Not Supported 00:15:18.689 Transport SGL Data Block: Not Supported 00:15:18.689 Replay Protected Memory Block: Not Supported 00:15:18.689 00:15:18.689 Firmware Slot Information 00:15:18.689 ========================= 00:15:18.689 Active slot: 1 00:15:18.689 Slot 1 Firmware Revision: 25.01 00:15:18.689 00:15:18.689 00:15:18.689 Commands Supported and Effects 00:15:18.689 ============================== 00:15:18.689 Admin Commands 00:15:18.689 -------------- 00:15:18.689 Get Log Page (02h): Supported 00:15:18.689 Identify (06h): Supported 00:15:18.689 Abort (08h): Supported 00:15:18.689 Set Features (09h): Supported 00:15:18.689 Get Features (0Ah): Supported 00:15:18.689 Asynchronous Event Request (0Ch): Supported 00:15:18.689 Keep Alive (18h): Supported 00:15:18.689 I/O Commands 00:15:18.689 ------------ 00:15:18.689 Flush (00h): Supported LBA-Change 00:15:18.689 Write (01h): Supported LBA-Change 00:15:18.689 Read (02h): Supported 00:15:18.689 Compare (05h): Supported 00:15:18.689 Write Zeroes (08h): Supported LBA-Change 00:15:18.689 Dataset Management (09h): Supported LBA-Change 00:15:18.689 Copy (19h): Supported LBA-Change 00:15:18.689 00:15:18.689 Error Log 00:15:18.689 ========= 00:15:18.689 00:15:18.689 Arbitration 00:15:18.689 =========== 00:15:18.689 Arbitration Burst: 1 00:15:18.689 00:15:18.689 Power Management 00:15:18.689 ================ 00:15:18.689 Number of Power States: 1 00:15:18.689 Current Power State: Power State #0 00:15:18.689 Power State #0: 00:15:18.689 Max Power: 0.00 W 00:15:18.689 Non-Operational State: Operational 00:15:18.689 Entry Latency: Not Reported 00:15:18.689 Exit Latency: Not Reported 00:15:18.689 Relative Read Throughput: 0 00:15:18.689 Relative Read Latency: 0 00:15:18.689 Relative Write Throughput: 0 00:15:18.689 Relative Write Latency: 0 00:15:18.689 Idle Power: Not Reported 00:15:18.689 Active Power: Not Reported 00:15:18.689 Non-Operational Permissive Mode: Not Supported 00:15:18.689 00:15:18.689 Health Information 00:15:18.689 ================== 00:15:18.689 Critical Warnings: 00:15:18.689 Available Spare Space: OK 00:15:18.689 Temperature: OK 00:15:18.689 Device Reliability: OK 00:15:18.689 Read Only: No 00:15:18.689 Volatile Memory Backup: OK 00:15:18.689 Current Temperature: 0 Kelvin (-273 Celsius) 00:15:18.689 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:15:18.689 Available Spare: 0% 00:15:18.689 Available Sp[2024-11-26 19:11:31.278314] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:15:18.689 [2024-11-26 19:11:31.278323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:15:18.689 [2024-11-26 19:11:31.278354] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Prepare to destruct SSD 00:15:18.689 [2024-11-26 19:11:31.278364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:18.689 [2024-11-26 19:11:31.278371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:18.689 [2024-11-26 19:11:31.278377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:18.689 [2024-11-26 19:11:31.278384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:18.689 [2024-11-26 19:11:31.280872] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:15:18.689 [2024-11-26 19:11:31.280884] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x464001 00:15:18.689 [2024-11-26 19:11:31.281368] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:18.689 [2024-11-26 19:11:31.281410] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] RTD3E = 0 us 00:15:18.689 [2024-11-26 19:11:31.281416] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] shutdown timeout = 10000 ms 00:15:18.689 [2024-11-26 19:11:31.282381] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x9 00:15:18.689 [2024-11-26 19:11:31.282393] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] shutdown complete in 0 milliseconds 00:15:18.689 [2024-11-26 19:11:31.282449] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user1/1/cntrl 00:15:18.689 [2024-11-26 19:11:31.285870] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:15:18.950 are Threshold: 0% 00:15:18.950 Life Percentage Used: 0% 00:15:18.950 Data Units Read: 0 00:15:18.950 Data Units Written: 0 00:15:18.950 Host Read Commands: 0 00:15:18.950 Host Write Commands: 0 00:15:18.950 Controller Busy Time: 0 minutes 00:15:18.950 Power Cycles: 0 00:15:18.950 Power On Hours: 0 hours 00:15:18.950 Unsafe Shutdowns: 0 00:15:18.950 Unrecoverable Media Errors: 0 00:15:18.950 Lifetime Error Log Entries: 0 00:15:18.950 Warning Temperature Time: 0 minutes 00:15:18.950 Critical Temperature Time: 0 minutes 00:15:18.950 00:15:18.950 Number of Queues 00:15:18.950 ================ 00:15:18.950 Number of I/O Submission Queues: 127 00:15:18.950 Number of I/O Completion Queues: 127 00:15:18.950 00:15:18.950 Active Namespaces 00:15:18.950 ================= 00:15:18.950 Namespace ID:1 00:15:18.950 Error Recovery Timeout: Unlimited 00:15:18.950 Command Set Identifier: NVM (00h) 00:15:18.950 Deallocate: Supported 00:15:18.950 Deallocated/Unwritten Error: Not Supported 00:15:18.950 Deallocated Read Value: Unknown 00:15:18.950 Deallocate in Write Zeroes: Not Supported 00:15:18.950 Deallocated Guard Field: 0xFFFF 00:15:18.950 Flush: Supported 00:15:18.950 Reservation: Supported 00:15:18.950 Namespace Sharing Capabilities: Multiple Controllers 00:15:18.950 Size (in LBAs): 131072 (0GiB) 00:15:18.950 Capacity (in LBAs): 131072 (0GiB) 00:15:18.950 Utilization (in LBAs): 131072 (0GiB) 00:15:18.950 NGUID: 96B405269B204EC3961FED682E967251 00:15:18.950 UUID: 96b40526-9b20-4ec3-961f-ed682e967251 00:15:18.950 Thin Provisioning: Not Supported 00:15:18.950 Per-NS Atomic Units: Yes 00:15:18.950 Atomic Boundary Size (Normal): 0 00:15:18.950 Atomic Boundary Size (PFail): 0 00:15:18.950 Atomic Boundary Offset: 0 00:15:18.950 Maximum Single Source Range Length: 65535 00:15:18.950 Maximum Copy Length: 65535 00:15:18.950 Maximum Source Range Count: 1 00:15:18.950 NGUID/EUI64 Never Reused: No 00:15:18.950 Namespace Write Protected: No 00:15:18.950 Number of LBA Formats: 1 00:15:18.950 Current LBA Format: LBA Format #00 00:15:18.950 LBA Format #00: Data Size: 512 Metadata Size: 0 00:15:18.950 00:15:18.950 19:11:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:15:18.950 [2024-11-26 19:11:31.490555] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:24.235 Initializing NVMe Controllers 00:15:24.235 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:15:24.235 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:15:24.235 Initialization complete. Launching workers. 00:15:24.235 ======================================================== 00:15:24.235 Latency(us) 00:15:24.235 Device Information : IOPS MiB/s Average min max 00:15:24.235 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 39940.58 156.02 3204.43 861.39 7771.36 00:15:24.235 ======================================================== 00:15:24.235 Total : 39940.58 156.02 3204.43 861.39 7771.36 00:15:24.235 00:15:24.235 [2024-11-26 19:11:36.508632] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:24.235 19:11:36 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:15:24.235 [2024-11-26 19:11:36.699531] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:29.520 Initializing NVMe Controllers 00:15:29.520 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:15:29.520 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:15:29.520 Initialization complete. Launching workers. 00:15:29.520 ======================================================== 00:15:29.520 Latency(us) 00:15:29.520 Device Information : IOPS MiB/s Average min max 00:15:29.520 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 16055.92 62.72 7977.68 5988.74 9969.43 00:15:29.520 ======================================================== 00:15:29.520 Total : 16055.92 62.72 7977.68 5988.74 9969.43 00:15:29.520 00:15:29.520 [2024-11-26 19:11:41.739431] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:29.520 19:11:41 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:15:29.520 [2024-11-26 19:11:41.946322] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:34.799 [2024-11-26 19:11:47.036138] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:34.799 Initializing NVMe Controllers 00:15:34.799 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:15:34.799 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:15:34.799 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 1 00:15:34.799 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 2 00:15:34.799 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 3 00:15:34.799 Initialization complete. Launching workers. 00:15:34.799 Starting thread on core 2 00:15:34.799 Starting thread on core 3 00:15:34.799 Starting thread on core 1 00:15:34.799 19:11:47 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -d 256 -g 00:15:34.799 [2024-11-26 19:11:47.325273] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:38.128 [2024-11-26 19:11:50.395584] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:38.128 Initializing NVMe Controllers 00:15:38.128 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:15:38.128 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:15:38.128 Associating SPDK bdev Controller (SPDK1 ) with lcore 0 00:15:38.128 Associating SPDK bdev Controller (SPDK1 ) with lcore 1 00:15:38.128 Associating SPDK bdev Controller (SPDK1 ) with lcore 2 00:15:38.128 Associating SPDK bdev Controller (SPDK1 ) with lcore 3 00:15:38.128 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:15:38.128 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:15:38.128 Initialization complete. Launching workers. 00:15:38.128 Starting thread on core 1 with urgent priority queue 00:15:38.128 Starting thread on core 2 with urgent priority queue 00:15:38.128 Starting thread on core 3 with urgent priority queue 00:15:38.128 Starting thread on core 0 with urgent priority queue 00:15:38.128 SPDK bdev Controller (SPDK1 ) core 0: 13374.67 IO/s 7.48 secs/100000 ios 00:15:38.128 SPDK bdev Controller (SPDK1 ) core 1: 13965.00 IO/s 7.16 secs/100000 ios 00:15:38.128 SPDK bdev Controller (SPDK1 ) core 2: 9925.33 IO/s 10.08 secs/100000 ios 00:15:38.128 SPDK bdev Controller (SPDK1 ) core 3: 11812.33 IO/s 8.47 secs/100000 ios 00:15:38.128 ======================================================== 00:15:38.128 00:15:38.128 19:11:50 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:15:38.128 [2024-11-26 19:11:50.690290] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:38.128 Initializing NVMe Controllers 00:15:38.128 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:15:38.128 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:15:38.128 Namespace ID: 1 size: 0GB 00:15:38.128 Initialization complete. 00:15:38.128 INFO: using host memory buffer for IO 00:15:38.128 Hello world! 00:15:38.128 [2024-11-26 19:11:50.724473] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:38.388 19:11:50 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:15:38.647 [2024-11-26 19:11:51.021289] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:39.587 Initializing NVMe Controllers 00:15:39.587 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:15:39.587 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:15:39.587 Initialization complete. Launching workers. 00:15:39.587 submit (in ns) avg, min, max = 8108.6, 3923.3, 4000367.5 00:15:39.587 complete (in ns) avg, min, max = 17803.3, 2366.7, 3998579.2 00:15:39.587 00:15:39.587 Submit histogram 00:15:39.587 ================ 00:15:39.587 Range in us Cumulative Count 00:15:39.587 3.920 - 3.947: 0.1691% ( 32) 00:15:39.587 3.947 - 3.973: 1.9610% ( 339) 00:15:39.587 3.973 - 4.000: 7.2890% ( 1008) 00:15:39.587 4.000 - 4.027: 16.5495% ( 1752) 00:15:39.587 4.027 - 4.053: 29.5470% ( 2459) 00:15:39.587 4.053 - 4.080: 43.3215% ( 2606) 00:15:39.587 4.080 - 4.107: 58.5390% ( 2879) 00:15:39.587 4.107 - 4.133: 75.0621% ( 3126) 00:15:39.587 4.133 - 4.160: 87.1928% ( 2295) 00:15:39.587 4.160 - 4.187: 94.5875% ( 1399) 00:15:39.587 4.187 - 4.213: 98.1712% ( 678) 00:15:39.587 4.213 - 4.240: 99.2706% ( 208) 00:15:39.587 4.240 - 4.267: 99.5084% ( 45) 00:15:39.587 4.267 - 4.293: 99.5401% ( 6) 00:15:39.587 4.293 - 4.320: 99.5560% ( 3) 00:15:39.587 4.347 - 4.373: 99.5613% ( 1) 00:15:39.587 4.373 - 4.400: 99.5666% ( 1) 00:15:39.587 4.507 - 4.533: 99.5719% ( 1) 00:15:39.587 4.587 - 4.613: 99.5771% ( 1) 00:15:39.587 4.880 - 4.907: 99.5824% ( 1) 00:15:39.587 4.907 - 4.933: 99.5877% ( 1) 00:15:39.587 4.933 - 4.960: 99.5930% ( 1) 00:15:39.587 4.960 - 4.987: 99.5983% ( 1) 00:15:39.587 5.147 - 5.173: 99.6036% ( 1) 00:15:39.587 5.200 - 5.227: 99.6089% ( 1) 00:15:39.587 5.280 - 5.307: 99.6141% ( 1) 00:15:39.587 5.413 - 5.440: 99.6194% ( 1) 00:15:39.587 5.440 - 5.467: 99.6247% ( 1) 00:15:39.587 5.760 - 5.787: 99.6300% ( 1) 00:15:39.587 5.867 - 5.893: 99.6353% ( 1) 00:15:39.587 5.893 - 5.920: 99.6406% ( 1) 00:15:39.587 6.027 - 6.053: 99.6459% ( 1) 00:15:39.587 6.053 - 6.080: 99.6511% ( 1) 00:15:39.587 6.080 - 6.107: 99.6617% ( 2) 00:15:39.587 6.107 - 6.133: 99.6776% ( 3) 00:15:39.587 6.133 - 6.160: 99.6881% ( 2) 00:15:39.587 6.160 - 6.187: 99.6934% ( 1) 00:15:39.587 6.187 - 6.213: 99.6987% ( 1) 00:15:39.587 6.213 - 6.240: 99.7040% ( 1) 00:15:39.587 6.240 - 6.267: 99.7146% ( 2) 00:15:39.587 6.267 - 6.293: 99.7199% ( 1) 00:15:39.587 6.293 - 6.320: 99.7251% ( 1) 00:15:39.587 6.347 - 6.373: 99.7357% ( 2) 00:15:39.587 6.373 - 6.400: 99.7463% ( 2) 00:15:39.587 6.400 - 6.427: 99.7516% ( 1) 00:15:39.587 6.453 - 6.480: 99.7569% ( 1) 00:15:39.587 6.480 - 6.507: 99.7621% ( 1) 00:15:39.587 6.507 - 6.533: 99.7674% ( 1) 00:15:39.587 6.533 - 6.560: 99.7727% ( 1) 00:15:39.587 6.560 - 6.587: 99.7886% ( 3) 00:15:39.587 6.720 - 6.747: 99.7939% ( 1) 00:15:39.587 6.747 - 6.773: 99.7991% ( 1) 00:15:39.587 6.800 - 6.827: 99.8097% ( 2) 00:15:39.587 6.880 - 6.933: 99.8150% ( 1) 00:15:39.587 7.093 - 7.147: 99.8256% ( 2) 00:15:39.587 7.200 - 7.253: 99.8361% ( 2) 00:15:39.587 7.253 - 7.307: 99.8467% ( 2) 00:15:39.587 7.360 - 7.413: 99.8520% ( 1) 00:15:39.587 7.467 - 7.520: 99.8573% ( 1) 00:15:39.587 7.520 - 7.573: 99.8626% ( 1) 00:15:39.587 7.627 - 7.680: 99.8679% ( 1) 00:15:39.587 7.787 - 7.840: 99.8731% ( 1) 00:15:39.587 7.840 - 7.893: 99.8837% ( 2) 00:15:39.587 8.320 - 8.373: 99.8890% ( 1) 00:15:39.587 8.907 - 8.960: 99.8943% ( 1) 00:15:39.587 10.987 - 11.040: 99.8996% ( 1) 00:15:39.587 3986.773 - 4014.080: 100.0000% ( 19) 00:15:39.587 00:15:39.587 Complete histogram 00:15:39.587 ================== 00:15:39.587 Range in us Cumulative Count 00:15:39.587 2.360 - 2.373: 0.0106% ( 2) 00:15:39.587 2.373 - 2.387: 0.0317% ( 4) 00:15:39.587 2.387 - 2.400: 1.0677% ( 196) 00:15:39.587 2.400 - 2.413: 1.2369% ( 32) 00:15:39.587 2.413 - 2.427: 1.3743% ( 26) 00:15:39.587 2.427 - 2.440: 1.4166% ( 8) 00:15:39.587 2.440 - 2.453: 13.1349% ( 2217) 00:15:39.587 2.453 - 2.467: 54.5536% ( 7836) 00:15:39.587 2.467 - 2.480: 66.0606% ( 2177) 00:15:39.587 2.480 - 2.493: 77.0072% ( 2071) 00:15:39.587 2.493 - [2024-11-26 19:11:52.044790] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:39.587 2.507: 80.3795% ( 638) 00:15:39.587 2.507 - 2.520: 81.9546% ( 298) 00:15:39.587 2.520 - 2.533: 86.9232% ( 940) 00:15:39.587 2.533 - 2.547: 93.2607% ( 1199) 00:15:39.587 2.547 - 2.560: 96.4903% ( 611) 00:15:39.587 2.560 - 2.573: 98.2240% ( 328) 00:15:39.587 2.573 - 2.587: 99.0644% ( 159) 00:15:39.587 2.587 - 2.600: 99.3023% ( 45) 00:15:39.587 2.600 - 2.613: 99.3816% ( 15) 00:15:39.587 2.613 - 2.627: 99.3921% ( 2) 00:15:39.587 2.787 - 2.800: 99.3974% ( 1) 00:15:39.587 4.347 - 4.373: 99.4027% ( 1) 00:15:39.587 4.400 - 4.427: 99.4133% ( 2) 00:15:39.587 4.427 - 4.453: 99.4186% ( 1) 00:15:39.587 4.507 - 4.533: 99.4291% ( 2) 00:15:39.587 4.560 - 4.587: 99.4344% ( 1) 00:15:39.587 4.587 - 4.613: 99.4397% ( 1) 00:15:39.587 4.613 - 4.640: 99.4450% ( 1) 00:15:39.587 4.640 - 4.667: 99.4503% ( 1) 00:15:39.587 4.667 - 4.693: 99.4556% ( 1) 00:15:39.587 4.693 - 4.720: 99.4609% ( 1) 00:15:39.587 4.747 - 4.773: 99.4714% ( 2) 00:15:39.587 4.773 - 4.800: 99.4767% ( 1) 00:15:39.587 4.800 - 4.827: 99.4873% ( 2) 00:15:39.587 4.827 - 4.853: 99.4926% ( 1) 00:15:39.587 4.853 - 4.880: 99.4979% ( 1) 00:15:39.587 4.907 - 4.933: 99.5031% ( 1) 00:15:39.587 4.933 - 4.960: 99.5137% ( 2) 00:15:39.587 4.960 - 4.987: 99.5243% ( 2) 00:15:39.588 4.987 - 5.013: 99.5296% ( 1) 00:15:39.588 5.093 - 5.120: 99.5349% ( 1) 00:15:39.588 5.120 - 5.147: 99.5401% ( 1) 00:15:39.588 5.200 - 5.227: 99.5507% ( 2) 00:15:39.588 5.307 - 5.333: 99.5560% ( 1) 00:15:39.588 5.333 - 5.360: 99.5613% ( 1) 00:15:39.588 5.440 - 5.467: 99.5666% ( 1) 00:15:39.588 5.467 - 5.493: 99.5719% ( 1) 00:15:39.588 5.520 - 5.547: 99.5771% ( 1) 00:15:39.588 5.573 - 5.600: 99.5824% ( 1) 00:15:39.588 6.320 - 6.347: 99.5877% ( 1) 00:15:39.588 6.373 - 6.400: 99.5930% ( 1) 00:15:39.588 6.880 - 6.933: 99.5983% ( 1) 00:15:39.588 11.093 - 11.147: 99.6036% ( 1) 00:15:39.588 14.293 - 14.400: 99.6089% ( 1) 00:15:39.588 43.947 - 44.160: 99.6141% ( 1) 00:15:39.588 2184.533 - 2198.187: 99.6194% ( 1) 00:15:39.588 3986.773 - 4014.080: 100.0000% ( 72) 00:15:39.588 00:15:39.588 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user1/1 nqn.2019-07.io.spdk:cnode1 1 00:15:39.588 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user1/1 00:15:39.588 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode1 00:15:39.588 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc3 00:15:39.588 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:15:39.848 [ 00:15:39.848 { 00:15:39.848 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:15:39.848 "subtype": "Discovery", 00:15:39.848 "listen_addresses": [], 00:15:39.848 "allow_any_host": true, 00:15:39.848 "hosts": [] 00:15:39.848 }, 00:15:39.848 { 00:15:39.848 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:15:39.848 "subtype": "NVMe", 00:15:39.848 "listen_addresses": [ 00:15:39.848 { 00:15:39.848 "trtype": "VFIOUSER", 00:15:39.848 "adrfam": "IPv4", 00:15:39.848 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:15:39.848 "trsvcid": "0" 00:15:39.848 } 00:15:39.848 ], 00:15:39.848 "allow_any_host": true, 00:15:39.848 "hosts": [], 00:15:39.848 "serial_number": "SPDK1", 00:15:39.848 "model_number": "SPDK bdev Controller", 00:15:39.848 "max_namespaces": 32, 00:15:39.848 "min_cntlid": 1, 00:15:39.848 "max_cntlid": 65519, 00:15:39.848 "namespaces": [ 00:15:39.848 { 00:15:39.848 "nsid": 1, 00:15:39.848 "bdev_name": "Malloc1", 00:15:39.848 "name": "Malloc1", 00:15:39.848 "nguid": "96B405269B204EC3961FED682E967251", 00:15:39.848 "uuid": "96b40526-9b20-4ec3-961f-ed682e967251" 00:15:39.848 } 00:15:39.848 ] 00:15:39.848 }, 00:15:39.848 { 00:15:39.848 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:15:39.848 "subtype": "NVMe", 00:15:39.848 "listen_addresses": [ 00:15:39.848 { 00:15:39.848 "trtype": "VFIOUSER", 00:15:39.848 "adrfam": "IPv4", 00:15:39.848 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:15:39.848 "trsvcid": "0" 00:15:39.848 } 00:15:39.848 ], 00:15:39.848 "allow_any_host": true, 00:15:39.848 "hosts": [], 00:15:39.848 "serial_number": "SPDK2", 00:15:39.848 "model_number": "SPDK bdev Controller", 00:15:39.848 "max_namespaces": 32, 00:15:39.848 "min_cntlid": 1, 00:15:39.848 "max_cntlid": 65519, 00:15:39.848 "namespaces": [ 00:15:39.848 { 00:15:39.848 "nsid": 1, 00:15:39.848 "bdev_name": "Malloc2", 00:15:39.848 "name": "Malloc2", 00:15:39.848 "nguid": "67897639123E44E1BF3F80023698BF86", 00:15:39.848 "uuid": "67897639-123e-44e1-bf3f-80023698bf86" 00:15:39.848 } 00:15:39.848 ] 00:15:39.848 } 00:15:39.848 ] 00:15:39.848 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:15:39.848 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=3810237 00:15:39.848 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -n 2 -g -t /tmp/aer_touch_file 00:15:39.848 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:15:39.848 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1269 -- # local i=0 00:15:39.849 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:15:39.849 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:15:39.849 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1280 -- # return 0 00:15:39.849 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:15:39.849 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc3 00:15:39.849 Malloc3 00:15:40.109 [2024-11-26 19:11:52.477890] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:15:40.109 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc3 -n 2 00:15:40.109 [2024-11-26 19:11:52.627966] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:15:40.109 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:15:40.109 Asynchronous Event Request test 00:15:40.109 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:15:40.109 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:15:40.109 Registering asynchronous event callbacks... 00:15:40.109 Starting namespace attribute notice tests for all controllers... 00:15:40.109 /var/run/vfio-user/domain/vfio-user1/1: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:15:40.109 aer_cb - Changed Namespace 00:15:40.109 Cleaning up... 00:15:40.372 [ 00:15:40.372 { 00:15:40.372 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:15:40.372 "subtype": "Discovery", 00:15:40.372 "listen_addresses": [], 00:15:40.372 "allow_any_host": true, 00:15:40.372 "hosts": [] 00:15:40.372 }, 00:15:40.372 { 00:15:40.372 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:15:40.372 "subtype": "NVMe", 00:15:40.372 "listen_addresses": [ 00:15:40.372 { 00:15:40.372 "trtype": "VFIOUSER", 00:15:40.372 "adrfam": "IPv4", 00:15:40.372 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:15:40.372 "trsvcid": "0" 00:15:40.372 } 00:15:40.372 ], 00:15:40.372 "allow_any_host": true, 00:15:40.372 "hosts": [], 00:15:40.372 "serial_number": "SPDK1", 00:15:40.372 "model_number": "SPDK bdev Controller", 00:15:40.372 "max_namespaces": 32, 00:15:40.372 "min_cntlid": 1, 00:15:40.372 "max_cntlid": 65519, 00:15:40.372 "namespaces": [ 00:15:40.372 { 00:15:40.372 "nsid": 1, 00:15:40.372 "bdev_name": "Malloc1", 00:15:40.372 "name": "Malloc1", 00:15:40.372 "nguid": "96B405269B204EC3961FED682E967251", 00:15:40.372 "uuid": "96b40526-9b20-4ec3-961f-ed682e967251" 00:15:40.372 }, 00:15:40.372 { 00:15:40.372 "nsid": 2, 00:15:40.372 "bdev_name": "Malloc3", 00:15:40.372 "name": "Malloc3", 00:15:40.372 "nguid": "7A771E4159474DA09551F7060D8371E1", 00:15:40.372 "uuid": "7a771e41-5947-4da0-9551-f7060d8371e1" 00:15:40.372 } 00:15:40.372 ] 00:15:40.372 }, 00:15:40.372 { 00:15:40.372 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:15:40.372 "subtype": "NVMe", 00:15:40.372 "listen_addresses": [ 00:15:40.372 { 00:15:40.372 "trtype": "VFIOUSER", 00:15:40.372 "adrfam": "IPv4", 00:15:40.372 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:15:40.372 "trsvcid": "0" 00:15:40.372 } 00:15:40.372 ], 00:15:40.372 "allow_any_host": true, 00:15:40.372 "hosts": [], 00:15:40.372 "serial_number": "SPDK2", 00:15:40.372 "model_number": "SPDK bdev Controller", 00:15:40.372 "max_namespaces": 32, 00:15:40.372 "min_cntlid": 1, 00:15:40.372 "max_cntlid": 65519, 00:15:40.372 "namespaces": [ 00:15:40.372 { 00:15:40.372 "nsid": 1, 00:15:40.372 "bdev_name": "Malloc2", 00:15:40.372 "name": "Malloc2", 00:15:40.372 "nguid": "67897639123E44E1BF3F80023698BF86", 00:15:40.372 "uuid": "67897639-123e-44e1-bf3f-80023698bf86" 00:15:40.372 } 00:15:40.372 ] 00:15:40.372 } 00:15:40.372 ] 00:15:40.372 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 3810237 00:15:40.372 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:15:40.372 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user2/2 00:15:40.372 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode2 00:15:40.372 19:11:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -L nvme -L nvme_vfio -L vfio_pci 00:15:40.372 [2024-11-26 19:11:52.866322] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:15:40.372 [2024-11-26 19:11:52.866364] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3810390 ] 00:15:40.372 [2024-11-26 19:11:52.918740] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user2/2 00:15:40.372 [2024-11-26 19:11:52.931102] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:15:40.372 [2024-11-26 19:11:52.931129] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7fe446f83000 00:15:40.372 [2024-11-26 19:11:52.932104] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.933111] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.934117] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.935130] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.936131] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.937139] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.938151] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.939155] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:15:40.372 [2024-11-26 19:11:52.940165] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:15:40.372 [2024-11-26 19:11:52.940175] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7fe446f78000 00:15:40.372 [2024-11-26 19:11:52.941502] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:15:40.372 [2024-11-26 19:11:52.957703] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user2/2/cntrl Setup Successfully 00:15:40.372 [2024-11-26 19:11:52.957731] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to connect adminq (no timeout) 00:15:40.372 [2024-11-26 19:11:52.959784] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:15:40.372 [2024-11-26 19:11:52.959833] nvme_pcie_common.c: 159:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:15:40.372 [2024-11-26 19:11:52.959927] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for connect adminq (no timeout) 00:15:40.372 [2024-11-26 19:11:52.959942] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read vs (no timeout) 00:15:40.372 [2024-11-26 19:11:52.959947] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read vs wait for vs (no timeout) 00:15:40.372 [2024-11-26 19:11:52.961867] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x8, value 0x10300 00:15:40.372 [2024-11-26 19:11:52.961880] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read cap (no timeout) 00:15:40.372 [2024-11-26 19:11:52.961888] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read cap wait for cap (no timeout) 00:15:40.372 [2024-11-26 19:11:52.962796] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:15:40.372 [2024-11-26 19:11:52.962805] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to check en (no timeout) 00:15:40.372 [2024-11-26 19:11:52.962813] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to check en wait for cc (timeout 15000 ms) 00:15:40.372 [2024-11-26 19:11:52.963801] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x0 00:15:40.372 [2024-11-26 19:11:52.963810] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:15:40.372 [2024-11-26 19:11:52.964804] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x0 00:15:40.372 [2024-11-26 19:11:52.964813] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CC.EN = 0 && CSTS.RDY = 0 00:15:40.372 [2024-11-26 19:11:52.964818] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to controller is disabled (timeout 15000 ms) 00:15:40.372 [2024-11-26 19:11:52.964826] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:15:40.372 [2024-11-26 19:11:52.964934] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Setting CC.EN = 1 00:15:40.372 [2024-11-26 19:11:52.964939] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:15:40.372 [2024-11-26 19:11:52.964944] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x28, value 0x2000003c0000 00:15:40.372 [2024-11-26 19:11:52.965824] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x30, value 0x2000003be000 00:15:40.372 [2024-11-26 19:11:52.966822] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x24, value 0xff00ff 00:15:40.372 [2024-11-26 19:11:52.967829] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:15:40.372 [2024-11-26 19:11:52.968831] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:15:40.372 [2024-11-26 19:11:52.968874] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:15:40.372 [2024-11-26 19:11:52.969836] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x1 00:15:40.372 [2024-11-26 19:11:52.969845] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:15:40.372 [2024-11-26 19:11:52.969853] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to reset admin queue (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.969877] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify controller (no timeout) 00:15:40.373 [2024-11-26 19:11:52.969885] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify controller (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.969901] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:15:40.373 [2024-11-26 19:11:52.969906] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:40.373 [2024-11-26 19:11:52.969910] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:40.373 [2024-11-26 19:11:52.969922] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:40.373 [2024-11-26 19:11:52.975871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:15:40.373 [2024-11-26 19:11:52.975883] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] transport max_xfer_size 131072 00:15:40.373 [2024-11-26 19:11:52.975889] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] MDTS max_xfer_size 131072 00:15:40.373 [2024-11-26 19:11:52.975893] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CNTLID 0x0001 00:15:40.373 [2024-11-26 19:11:52.975898] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:15:40.373 [2024-11-26 19:11:52.975903] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] transport max_sges 1 00:15:40.373 [2024-11-26 19:11:52.975908] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] fuses compare and write: 1 00:15:40.373 [2024-11-26 19:11:52.975913] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to configure AER (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.975921] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for configure aer (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.975931] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:15:40.373 [2024-11-26 19:11:52.983869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:15:40.373 [2024-11-26 19:11:52.983881] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:15:40.373 [2024-11-26 19:11:52.983890] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:15:40.373 [2024-11-26 19:11:52.983899] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:15:40.373 [2024-11-26 19:11:52.983907] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:15:40.373 [2024-11-26 19:11:52.983912] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set keep alive timeout (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.983922] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.983931] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:15:40.373 [2024-11-26 19:11:52.991868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:15:40.373 [2024-11-26 19:11:52.991876] nvme_ctrlr.c:3047:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Controller adjusted keep alive timeout to 0 ms 00:15:40.373 [2024-11-26 19:11:52.991881] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify controller iocs specific (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.991890] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set number of queues (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.991896] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for set number of queues (timeout 30000 ms) 00:15:40.373 [2024-11-26 19:11:52.991905] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:15:40.635 [2024-11-26 19:11:52.999867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:15:40.635 [2024-11-26 19:11:52.999935] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify active ns (timeout 30000 ms) 00:15:40.635 [2024-11-26 19:11:52.999943] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify active ns (timeout 30000 ms) 00:15:40.635 [2024-11-26 19:11:52.999951] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:15:40.635 [2024-11-26 19:11:52.999956] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:15:40.635 [2024-11-26 19:11:52.999960] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:40.635 [2024-11-26 19:11:52.999966] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:15:40.635 [2024-11-26 19:11:53.007868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:15:40.635 [2024-11-26 19:11:53.007881] nvme_ctrlr.c:4735:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Namespace 1 was added 00:15:40.635 [2024-11-26 19:11:53.007894] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify ns (timeout 30000 ms) 00:15:40.635 [2024-11-26 19:11:53.007902] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify ns (timeout 30000 ms) 00:15:40.635 [2024-11-26 19:11:53.007909] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:15:40.635 [2024-11-26 19:11:53.007914] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:40.635 [2024-11-26 19:11:53.007918] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:40.635 [2024-11-26 19:11:53.007924] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.015869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.015881] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify namespace id descriptors (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.015890] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.015897] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:15:40.636 [2024-11-26 19:11:53.015905] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:40.636 [2024-11-26 19:11:53.015908] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:40.636 [2024-11-26 19:11:53.015915] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.023866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.023878] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify ns iocs specific (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.023886] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set supported log pages (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.023893] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set supported features (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.023899] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set host behavior support feature (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.023905] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set doorbell buffer config (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.023910] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set host ID (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.023915] nvme_ctrlr.c:3147:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] NVMe-oF transport - not sending Set Features - Host ID 00:15:40.636 [2024-11-26 19:11:53.023919] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to transport ready (timeout 30000 ms) 00:15:40.636 [2024-11-26 19:11:53.023924] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to ready (no timeout) 00:15:40.636 [2024-11-26 19:11:53.023941] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.031867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.031880] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.039867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.039880] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.047868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.047881] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.055868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.055884] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:15:40.636 [2024-11-26 19:11:53.055889] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:15:40.636 [2024-11-26 19:11:53.055893] nvme_pcie_common.c:1275:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:15:40.636 [2024-11-26 19:11:53.055896] nvme_pcie_common.c:1291:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:15:40.636 [2024-11-26 19:11:53.055900] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:15:40.636 [2024-11-26 19:11:53.055906] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:15:40.636 [2024-11-26 19:11:53.055916] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:15:40.636 [2024-11-26 19:11:53.055921] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:15:40.636 [2024-11-26 19:11:53.055924] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:40.636 [2024-11-26 19:11:53.055930] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.055938] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:15:40.636 [2024-11-26 19:11:53.055942] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:15:40.636 [2024-11-26 19:11:53.055945] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:40.636 [2024-11-26 19:11:53.055951] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.055959] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:15:40.636 [2024-11-26 19:11:53.055964] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:15:40.636 [2024-11-26 19:11:53.055967] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:15:40.636 [2024-11-26 19:11:53.055973] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:15:40.636 [2024-11-26 19:11:53.063869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.063884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.063895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:15:40.636 [2024-11-26 19:11:53.063902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:15:40.636 ===================================================== 00:15:40.636 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:15:40.636 ===================================================== 00:15:40.636 Controller Capabilities/Features 00:15:40.636 ================================ 00:15:40.636 Vendor ID: 4e58 00:15:40.636 Subsystem Vendor ID: 4e58 00:15:40.636 Serial Number: SPDK2 00:15:40.636 Model Number: SPDK bdev Controller 00:15:40.636 Firmware Version: 25.01 00:15:40.636 Recommended Arb Burst: 6 00:15:40.636 IEEE OUI Identifier: 8d 6b 50 00:15:40.636 Multi-path I/O 00:15:40.636 May have multiple subsystem ports: Yes 00:15:40.636 May have multiple controllers: Yes 00:15:40.636 Associated with SR-IOV VF: No 00:15:40.636 Max Data Transfer Size: 131072 00:15:40.636 Max Number of Namespaces: 32 00:15:40.636 Max Number of I/O Queues: 127 00:15:40.636 NVMe Specification Version (VS): 1.3 00:15:40.636 NVMe Specification Version (Identify): 1.3 00:15:40.636 Maximum Queue Entries: 256 00:15:40.636 Contiguous Queues Required: Yes 00:15:40.636 Arbitration Mechanisms Supported 00:15:40.636 Weighted Round Robin: Not Supported 00:15:40.636 Vendor Specific: Not Supported 00:15:40.636 Reset Timeout: 15000 ms 00:15:40.636 Doorbell Stride: 4 bytes 00:15:40.636 NVM Subsystem Reset: Not Supported 00:15:40.636 Command Sets Supported 00:15:40.636 NVM Command Set: Supported 00:15:40.636 Boot Partition: Not Supported 00:15:40.636 Memory Page Size Minimum: 4096 bytes 00:15:40.636 Memory Page Size Maximum: 4096 bytes 00:15:40.636 Persistent Memory Region: Not Supported 00:15:40.636 Optional Asynchronous Events Supported 00:15:40.636 Namespace Attribute Notices: Supported 00:15:40.636 Firmware Activation Notices: Not Supported 00:15:40.636 ANA Change Notices: Not Supported 00:15:40.636 PLE Aggregate Log Change Notices: Not Supported 00:15:40.636 LBA Status Info Alert Notices: Not Supported 00:15:40.636 EGE Aggregate Log Change Notices: Not Supported 00:15:40.636 Normal NVM Subsystem Shutdown event: Not Supported 00:15:40.636 Zone Descriptor Change Notices: Not Supported 00:15:40.636 Discovery Log Change Notices: Not Supported 00:15:40.636 Controller Attributes 00:15:40.636 128-bit Host Identifier: Supported 00:15:40.636 Non-Operational Permissive Mode: Not Supported 00:15:40.636 NVM Sets: Not Supported 00:15:40.636 Read Recovery Levels: Not Supported 00:15:40.636 Endurance Groups: Not Supported 00:15:40.636 Predictable Latency Mode: Not Supported 00:15:40.636 Traffic Based Keep ALive: Not Supported 00:15:40.636 Namespace Granularity: Not Supported 00:15:40.636 SQ Associations: Not Supported 00:15:40.636 UUID List: Not Supported 00:15:40.636 Multi-Domain Subsystem: Not Supported 00:15:40.636 Fixed Capacity Management: Not Supported 00:15:40.636 Variable Capacity Management: Not Supported 00:15:40.636 Delete Endurance Group: Not Supported 00:15:40.636 Delete NVM Set: Not Supported 00:15:40.636 Extended LBA Formats Supported: Not Supported 00:15:40.636 Flexible Data Placement Supported: Not Supported 00:15:40.636 00:15:40.636 Controller Memory Buffer Support 00:15:40.636 ================================ 00:15:40.636 Supported: No 00:15:40.636 00:15:40.636 Persistent Memory Region Support 00:15:40.636 ================================ 00:15:40.636 Supported: No 00:15:40.636 00:15:40.636 Admin Command Set Attributes 00:15:40.636 ============================ 00:15:40.636 Security Send/Receive: Not Supported 00:15:40.636 Format NVM: Not Supported 00:15:40.636 Firmware Activate/Download: Not Supported 00:15:40.636 Namespace Management: Not Supported 00:15:40.636 Device Self-Test: Not Supported 00:15:40.637 Directives: Not Supported 00:15:40.637 NVMe-MI: Not Supported 00:15:40.637 Virtualization Management: Not Supported 00:15:40.637 Doorbell Buffer Config: Not Supported 00:15:40.637 Get LBA Status Capability: Not Supported 00:15:40.637 Command & Feature Lockdown Capability: Not Supported 00:15:40.637 Abort Command Limit: 4 00:15:40.637 Async Event Request Limit: 4 00:15:40.637 Number of Firmware Slots: N/A 00:15:40.637 Firmware Slot 1 Read-Only: N/A 00:15:40.637 Firmware Activation Without Reset: N/A 00:15:40.637 Multiple Update Detection Support: N/A 00:15:40.637 Firmware Update Granularity: No Information Provided 00:15:40.637 Per-Namespace SMART Log: No 00:15:40.637 Asymmetric Namespace Access Log Page: Not Supported 00:15:40.637 Subsystem NQN: nqn.2019-07.io.spdk:cnode2 00:15:40.637 Command Effects Log Page: Supported 00:15:40.637 Get Log Page Extended Data: Supported 00:15:40.637 Telemetry Log Pages: Not Supported 00:15:40.637 Persistent Event Log Pages: Not Supported 00:15:40.637 Supported Log Pages Log Page: May Support 00:15:40.637 Commands Supported & Effects Log Page: Not Supported 00:15:40.637 Feature Identifiers & Effects Log Page:May Support 00:15:40.637 NVMe-MI Commands & Effects Log Page: May Support 00:15:40.637 Data Area 4 for Telemetry Log: Not Supported 00:15:40.637 Error Log Page Entries Supported: 128 00:15:40.637 Keep Alive: Supported 00:15:40.637 Keep Alive Granularity: 10000 ms 00:15:40.637 00:15:40.637 NVM Command Set Attributes 00:15:40.637 ========================== 00:15:40.637 Submission Queue Entry Size 00:15:40.637 Max: 64 00:15:40.637 Min: 64 00:15:40.637 Completion Queue Entry Size 00:15:40.637 Max: 16 00:15:40.637 Min: 16 00:15:40.637 Number of Namespaces: 32 00:15:40.637 Compare Command: Supported 00:15:40.637 Write Uncorrectable Command: Not Supported 00:15:40.637 Dataset Management Command: Supported 00:15:40.637 Write Zeroes Command: Supported 00:15:40.637 Set Features Save Field: Not Supported 00:15:40.637 Reservations: Not Supported 00:15:40.637 Timestamp: Not Supported 00:15:40.637 Copy: Supported 00:15:40.637 Volatile Write Cache: Present 00:15:40.637 Atomic Write Unit (Normal): 1 00:15:40.637 Atomic Write Unit (PFail): 1 00:15:40.637 Atomic Compare & Write Unit: 1 00:15:40.637 Fused Compare & Write: Supported 00:15:40.637 Scatter-Gather List 00:15:40.637 SGL Command Set: Supported (Dword aligned) 00:15:40.637 SGL Keyed: Not Supported 00:15:40.637 SGL Bit Bucket Descriptor: Not Supported 00:15:40.637 SGL Metadata Pointer: Not Supported 00:15:40.637 Oversized SGL: Not Supported 00:15:40.637 SGL Metadata Address: Not Supported 00:15:40.637 SGL Offset: Not Supported 00:15:40.637 Transport SGL Data Block: Not Supported 00:15:40.637 Replay Protected Memory Block: Not Supported 00:15:40.637 00:15:40.637 Firmware Slot Information 00:15:40.637 ========================= 00:15:40.637 Active slot: 1 00:15:40.637 Slot 1 Firmware Revision: 25.01 00:15:40.637 00:15:40.637 00:15:40.637 Commands Supported and Effects 00:15:40.637 ============================== 00:15:40.637 Admin Commands 00:15:40.637 -------------- 00:15:40.637 Get Log Page (02h): Supported 00:15:40.637 Identify (06h): Supported 00:15:40.637 Abort (08h): Supported 00:15:40.637 Set Features (09h): Supported 00:15:40.637 Get Features (0Ah): Supported 00:15:40.637 Asynchronous Event Request (0Ch): Supported 00:15:40.637 Keep Alive (18h): Supported 00:15:40.637 I/O Commands 00:15:40.637 ------------ 00:15:40.637 Flush (00h): Supported LBA-Change 00:15:40.637 Write (01h): Supported LBA-Change 00:15:40.637 Read (02h): Supported 00:15:40.637 Compare (05h): Supported 00:15:40.637 Write Zeroes (08h): Supported LBA-Change 00:15:40.637 Dataset Management (09h): Supported LBA-Change 00:15:40.637 Copy (19h): Supported LBA-Change 00:15:40.637 00:15:40.637 Error Log 00:15:40.637 ========= 00:15:40.637 00:15:40.637 Arbitration 00:15:40.637 =========== 00:15:40.637 Arbitration Burst: 1 00:15:40.637 00:15:40.637 Power Management 00:15:40.637 ================ 00:15:40.637 Number of Power States: 1 00:15:40.637 Current Power State: Power State #0 00:15:40.637 Power State #0: 00:15:40.637 Max Power: 0.00 W 00:15:40.637 Non-Operational State: Operational 00:15:40.637 Entry Latency: Not Reported 00:15:40.637 Exit Latency: Not Reported 00:15:40.637 Relative Read Throughput: 0 00:15:40.637 Relative Read Latency: 0 00:15:40.637 Relative Write Throughput: 0 00:15:40.637 Relative Write Latency: 0 00:15:40.637 Idle Power: Not Reported 00:15:40.637 Active Power: Not Reported 00:15:40.637 Non-Operational Permissive Mode: Not Supported 00:15:40.637 00:15:40.637 Health Information 00:15:40.637 ================== 00:15:40.637 Critical Warnings: 00:15:40.637 Available Spare Space: OK 00:15:40.637 Temperature: OK 00:15:40.637 Device Reliability: OK 00:15:40.637 Read Only: No 00:15:40.637 Volatile Memory Backup: OK 00:15:40.637 Current Temperature: 0 Kelvin (-273 Celsius) 00:15:40.637 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:15:40.637 Available Spare: 0% 00:15:40.637 Available Sp[2024-11-26 19:11:53.064004] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:15:40.637 [2024-11-26 19:11:53.071868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:15:40.637 [2024-11-26 19:11:53.071901] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Prepare to destruct SSD 00:15:40.637 [2024-11-26 19:11:53.071911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:40.637 [2024-11-26 19:11:53.071918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:40.637 [2024-11-26 19:11:53.071924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:40.637 [2024-11-26 19:11:53.071931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:40.637 [2024-11-26 19:11:53.071972] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:15:40.637 [2024-11-26 19:11:53.071982] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x464001 00:15:40.637 [2024-11-26 19:11:53.072983] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:15:40.637 [2024-11-26 19:11:53.073034] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] RTD3E = 0 us 00:15:40.637 [2024-11-26 19:11:53.073045] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] shutdown timeout = 10000 ms 00:15:40.637 [2024-11-26 19:11:53.073988] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x9 00:15:40.637 [2024-11-26 19:11:53.074001] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] shutdown complete in 0 milliseconds 00:15:40.637 [2024-11-26 19:11:53.074050] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user2/2/cntrl 00:15:40.637 [2024-11-26 19:11:53.076868] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:15:40.637 are Threshold: 0% 00:15:40.637 Life Percentage Used: 0% 00:15:40.637 Data Units Read: 0 00:15:40.637 Data Units Written: 0 00:15:40.637 Host Read Commands: 0 00:15:40.637 Host Write Commands: 0 00:15:40.637 Controller Busy Time: 0 minutes 00:15:40.637 Power Cycles: 0 00:15:40.637 Power On Hours: 0 hours 00:15:40.637 Unsafe Shutdowns: 0 00:15:40.637 Unrecoverable Media Errors: 0 00:15:40.637 Lifetime Error Log Entries: 0 00:15:40.638 Warning Temperature Time: 0 minutes 00:15:40.638 Critical Temperature Time: 0 minutes 00:15:40.638 00:15:40.638 Number of Queues 00:15:40.638 ================ 00:15:40.638 Number of I/O Submission Queues: 127 00:15:40.638 Number of I/O Completion Queues: 127 00:15:40.638 00:15:40.638 Active Namespaces 00:15:40.638 ================= 00:15:40.638 Namespace ID:1 00:15:40.638 Error Recovery Timeout: Unlimited 00:15:40.638 Command Set Identifier: NVM (00h) 00:15:40.638 Deallocate: Supported 00:15:40.638 Deallocated/Unwritten Error: Not Supported 00:15:40.638 Deallocated Read Value: Unknown 00:15:40.638 Deallocate in Write Zeroes: Not Supported 00:15:40.638 Deallocated Guard Field: 0xFFFF 00:15:40.638 Flush: Supported 00:15:40.638 Reservation: Supported 00:15:40.638 Namespace Sharing Capabilities: Multiple Controllers 00:15:40.638 Size (in LBAs): 131072 (0GiB) 00:15:40.638 Capacity (in LBAs): 131072 (0GiB) 00:15:40.638 Utilization (in LBAs): 131072 (0GiB) 00:15:40.638 NGUID: 67897639123E44E1BF3F80023698BF86 00:15:40.638 UUID: 67897639-123e-44e1-bf3f-80023698bf86 00:15:40.638 Thin Provisioning: Not Supported 00:15:40.638 Per-NS Atomic Units: Yes 00:15:40.638 Atomic Boundary Size (Normal): 0 00:15:40.638 Atomic Boundary Size (PFail): 0 00:15:40.638 Atomic Boundary Offset: 0 00:15:40.638 Maximum Single Source Range Length: 65535 00:15:40.638 Maximum Copy Length: 65535 00:15:40.638 Maximum Source Range Count: 1 00:15:40.638 NGUID/EUI64 Never Reused: No 00:15:40.638 Namespace Write Protected: No 00:15:40.638 Number of LBA Formats: 1 00:15:40.638 Current LBA Format: LBA Format #00 00:15:40.638 LBA Format #00: Data Size: 512 Metadata Size: 0 00:15:40.638 00:15:40.638 19:11:53 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:15:40.898 [2024-11-26 19:11:53.281214] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:15:46.181 Initializing NVMe Controllers 00:15:46.181 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:15:46.181 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:15:46.181 Initialization complete. Launching workers. 00:15:46.181 ======================================================== 00:15:46.181 Latency(us) 00:15:46.181 Device Information : IOPS MiB/s Average min max 00:15:46.181 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 39982.40 156.18 3201.64 856.94 7777.52 00:15:46.181 ======================================================== 00:15:46.181 Total : 39982.40 156.18 3201.64 856.94 7777.52 00:15:46.181 00:15:46.181 [2024-11-26 19:11:58.381061] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:15:46.181 19:11:58 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:15:46.181 [2024-11-26 19:11:58.571794] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:15:51.469 Initializing NVMe Controllers 00:15:51.469 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:15:51.469 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:15:51.469 Initialization complete. Launching workers. 00:15:51.469 ======================================================== 00:15:51.469 Latency(us) 00:15:51.469 Device Information : IOPS MiB/s Average min max 00:15:51.469 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 34309.40 134.02 3732.17 1122.04 8986.18 00:15:51.469 ======================================================== 00:15:51.469 Total : 34309.40 134.02 3732.17 1122.04 8986.18 00:15:51.469 00:15:51.469 [2024-11-26 19:12:03.592495] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:15:51.469 19:12:03 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:15:51.469 [2024-11-26 19:12:03.804240] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:15:56.762 [2024-11-26 19:12:08.948947] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:15:56.762 Initializing NVMe Controllers 00:15:56.762 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:15:56.762 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:15:56.762 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 1 00:15:56.762 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 2 00:15:56.762 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 3 00:15:56.762 Initialization complete. Launching workers. 00:15:56.762 Starting thread on core 2 00:15:56.762 Starting thread on core 3 00:15:56.762 Starting thread on core 1 00:15:56.762 19:12:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -d 256 -g 00:15:56.762 [2024-11-26 19:12:09.242266] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:16:00.064 [2024-11-26 19:12:12.295122] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:16:00.064 Initializing NVMe Controllers 00:16:00.064 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:16:00.064 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:16:00.064 Associating SPDK bdev Controller (SPDK2 ) with lcore 0 00:16:00.064 Associating SPDK bdev Controller (SPDK2 ) with lcore 1 00:16:00.064 Associating SPDK bdev Controller (SPDK2 ) with lcore 2 00:16:00.064 Associating SPDK bdev Controller (SPDK2 ) with lcore 3 00:16:00.064 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:16:00.064 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:16:00.064 Initialization complete. Launching workers. 00:16:00.064 Starting thread on core 1 with urgent priority queue 00:16:00.064 Starting thread on core 2 with urgent priority queue 00:16:00.064 Starting thread on core 3 with urgent priority queue 00:16:00.064 Starting thread on core 0 with urgent priority queue 00:16:00.064 SPDK bdev Controller (SPDK2 ) core 0: 14202.33 IO/s 7.04 secs/100000 ios 00:16:00.064 SPDK bdev Controller (SPDK2 ) core 1: 12435.00 IO/s 8.04 secs/100000 ios 00:16:00.064 SPDK bdev Controller (SPDK2 ) core 2: 10494.67 IO/s 9.53 secs/100000 ios 00:16:00.064 SPDK bdev Controller (SPDK2 ) core 3: 9834.67 IO/s 10.17 secs/100000 ios 00:16:00.064 ======================================================== 00:16:00.064 00:16:00.064 19:12:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:16:00.064 [2024-11-26 19:12:12.587669] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:16:00.064 Initializing NVMe Controllers 00:16:00.064 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:16:00.064 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:16:00.064 Namespace ID: 1 size: 0GB 00:16:00.064 Initialization complete. 00:16:00.064 INFO: using host memory buffer for IO 00:16:00.064 Hello world! 00:16:00.064 [2024-11-26 19:12:12.597730] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:16:00.064 19:12:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:16:00.324 [2024-11-26 19:12:12.894828] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:16:01.711 Initializing NVMe Controllers 00:16:01.711 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:16:01.711 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:16:01.711 Initialization complete. Launching workers. 00:16:01.711 submit (in ns) avg, min, max = 8049.5, 3893.3, 4006034.2 00:16:01.711 complete (in ns) avg, min, max = 19284.0, 2395.8, 4035080.0 00:16:01.711 00:16:01.711 Submit histogram 00:16:01.711 ================ 00:16:01.711 Range in us Cumulative Count 00:16:01.711 3.893 - 3.920: 1.2464% ( 237) 00:16:01.711 3.920 - 3.947: 5.7852% ( 863) 00:16:01.711 3.947 - 3.973: 13.5374% ( 1474) 00:16:01.711 3.973 - 4.000: 24.2558% ( 2038) 00:16:01.711 4.000 - 4.027: 36.7782% ( 2381) 00:16:01.711 4.027 - 4.053: 51.6356% ( 2825) 00:16:01.711 4.053 - 4.080: 68.6915% ( 3243) 00:16:01.711 4.080 - 4.107: 83.1861% ( 2756) 00:16:01.711 4.107 - 4.133: 92.5055% ( 1772) 00:16:01.711 4.133 - 4.160: 97.1390% ( 881) 00:16:01.711 4.160 - 4.187: 98.8430% ( 324) 00:16:01.711 4.187 - 4.213: 99.3058% ( 88) 00:16:01.711 4.213 - 4.240: 99.4267% ( 23) 00:16:01.711 4.240 - 4.267: 99.4530% ( 5) 00:16:01.711 4.267 - 4.293: 99.4636% ( 2) 00:16:01.711 4.293 - 4.320: 99.4688% ( 1) 00:16:01.711 4.560 - 4.587: 99.4741% ( 1) 00:16:01.711 4.720 - 4.747: 99.4793% ( 1) 00:16:01.711 4.747 - 4.773: 99.4846% ( 1) 00:16:01.711 4.853 - 4.880: 99.4898% ( 1) 00:16:01.711 5.067 - 5.093: 99.4951% ( 1) 00:16:01.712 5.120 - 5.147: 99.5004% ( 1) 00:16:01.712 5.200 - 5.227: 99.5056% ( 1) 00:16:01.712 5.280 - 5.307: 99.5161% ( 2) 00:16:01.712 5.413 - 5.440: 99.5267% ( 2) 00:16:01.712 5.440 - 5.467: 99.5319% ( 1) 00:16:01.712 5.573 - 5.600: 99.5372% ( 1) 00:16:01.712 5.600 - 5.627: 99.5424% ( 1) 00:16:01.712 5.813 - 5.840: 99.5477% ( 1) 00:16:01.712 5.973 - 6.000: 99.5530% ( 1) 00:16:01.712 6.000 - 6.027: 99.5582% ( 1) 00:16:01.712 6.027 - 6.053: 99.5687% ( 2) 00:16:01.712 6.053 - 6.080: 99.5793% ( 2) 00:16:01.712 6.107 - 6.133: 99.5845% ( 1) 00:16:01.712 6.160 - 6.187: 99.5898% ( 1) 00:16:01.712 6.187 - 6.213: 99.5950% ( 1) 00:16:01.712 6.267 - 6.293: 99.6003% ( 1) 00:16:01.712 6.293 - 6.320: 99.6161% ( 3) 00:16:01.712 6.347 - 6.373: 99.6319% ( 3) 00:16:01.712 6.400 - 6.427: 99.6424% ( 2) 00:16:01.712 6.427 - 6.453: 99.6476% ( 1) 00:16:01.712 6.480 - 6.507: 99.6529% ( 1) 00:16:01.712 6.507 - 6.533: 99.6581% ( 1) 00:16:01.712 6.533 - 6.560: 99.6739% ( 3) 00:16:01.712 6.587 - 6.613: 99.6844% ( 2) 00:16:01.712 6.613 - 6.640: 99.6950% ( 2) 00:16:01.712 6.640 - 6.667: 99.7055% ( 2) 00:16:01.712 6.667 - 6.693: 99.7160% ( 2) 00:16:01.712 6.693 - 6.720: 99.7265% ( 2) 00:16:01.712 6.773 - 6.800: 99.7370% ( 2) 00:16:01.712 6.827 - 6.880: 99.7528% ( 3) 00:16:01.712 6.880 - 6.933: 99.7633% ( 2) 00:16:01.712 6.933 - 6.987: 99.7739% ( 2) 00:16:01.712 6.987 - 7.040: 99.7844% ( 2) 00:16:01.712 7.040 - 7.093: 99.8001% ( 3) 00:16:01.712 7.093 - 7.147: 99.8054% ( 1) 00:16:01.712 7.147 - 7.200: 99.8212% ( 3) 00:16:01.712 7.200 - 7.253: 99.8264% ( 1) 00:16:01.712 7.253 - 7.307: 99.8370% ( 2) 00:16:01.712 7.307 - 7.360: 99.8475% ( 2) 00:16:01.712 7.360 - 7.413: 99.8527% ( 1) 00:16:01.712 7.413 - 7.467: 99.8580% ( 1) 00:16:01.712 7.467 - 7.520: 99.8633% ( 1) 00:16:01.712 7.520 - 7.573: 99.8685% ( 1) 00:16:01.712 7.787 - 7.840: 99.8790% ( 2) 00:16:01.712 7.947 - 8.000: 99.8843% ( 1) 00:16:01.712 8.213 - 8.267: 99.8948% ( 2) 00:16:01.712 9.227 - 9.280: 99.9001% ( 1) 00:16:01.712 3986.773 - 4014.080: 100.0000% ( 19) 00:16:01.712 00:16:01.712 Complete histogram 00:16:01.712 ================== 00:16:01.712 Range in us Cumulative Count 00:16:01.712 2.387 - 2.400: 0.2367% ( 45) 00:16:01.712 2.400 - 2.413: 0.9519% ( 136) 00:16:01.712 2.413 - 2.427: 1.0413% ( 17) 00:16:01.712 2.427 - 2.440: 1.7829% ( 141) 00:16:01.712 2.440 - 2.453: 56.7634% ( 10454) 00:16:01.712 2.453 - 2.467: 63.4795% ( 1277) 00:16:01.712 2.467 - 2.480: 76.9328% ( 2558) 00:16:01.712 2.480 - 2.493: 79.8990% ( 564) 00:16:01.712 2.493 - [2024-11-26 19:12:13.987534] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:16:01.712 2.507: 81.3085% ( 268) 00:16:01.712 2.507 - 2.520: 85.7579% ( 846) 00:16:01.712 2.520 - 2.533: 92.1532% ( 1216) 00:16:01.712 2.533 - 2.547: 96.2186% ( 773) 00:16:01.712 2.547 - 2.560: 97.9962% ( 338) 00:16:01.712 2.560 - 2.573: 98.8903% ( 170) 00:16:01.712 2.573 - 2.587: 99.2953% ( 77) 00:16:01.712 2.587 - 2.600: 99.3899% ( 18) 00:16:01.712 2.600 - 2.613: 99.4110% ( 4) 00:16:01.712 2.613 - 2.627: 99.4215% ( 2) 00:16:01.712 2.653 - 2.667: 99.4267% ( 1) 00:16:01.712 4.693 - 4.720: 99.4320% ( 1) 00:16:01.712 4.800 - 4.827: 99.4373% ( 1) 00:16:01.712 4.853 - 4.880: 99.4425% ( 1) 00:16:01.712 4.880 - 4.907: 99.4478% ( 1) 00:16:01.712 4.907 - 4.933: 99.4583% ( 2) 00:16:01.712 4.933 - 4.960: 99.4636% ( 1) 00:16:01.712 4.960 - 4.987: 99.4688% ( 1) 00:16:01.712 4.987 - 5.013: 99.4741% ( 1) 00:16:01.712 5.067 - 5.093: 99.4793% ( 1) 00:16:01.712 5.093 - 5.120: 99.4846% ( 1) 00:16:01.712 5.173 - 5.200: 99.4898% ( 1) 00:16:01.712 5.200 - 5.227: 99.4951% ( 1) 00:16:01.712 5.280 - 5.307: 99.5004% ( 1) 00:16:01.712 5.333 - 5.360: 99.5056% ( 1) 00:16:01.712 5.467 - 5.493: 99.5109% ( 1) 00:16:01.712 5.493 - 5.520: 99.5161% ( 1) 00:16:01.712 5.520 - 5.547: 99.5214% ( 1) 00:16:01.712 5.600 - 5.627: 99.5267% ( 1) 00:16:01.712 5.707 - 5.733: 99.5319% ( 1) 00:16:01.712 5.813 - 5.840: 99.5372% ( 1) 00:16:01.712 5.840 - 5.867: 99.5477% ( 2) 00:16:01.712 6.027 - 6.053: 99.5530% ( 1) 00:16:01.712 6.080 - 6.107: 99.5582% ( 1) 00:16:01.712 6.133 - 6.160: 99.5635% ( 1) 00:16:01.712 6.453 - 6.480: 99.5687% ( 1) 00:16:01.712 6.560 - 6.587: 99.5740% ( 1) 00:16:01.712 44.587 - 44.800: 99.5793% ( 1) 00:16:01.712 3986.773 - 4014.080: 99.9947% ( 79) 00:16:01.712 4014.080 - 4041.387: 100.0000% ( 1) 00:16:01.712 00:16:01.712 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user2/2 nqn.2019-07.io.spdk:cnode2 2 00:16:01.712 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user2/2 00:16:01.712 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode2 00:16:01.712 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc4 00:16:01.712 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:16:01.712 [ 00:16:01.712 { 00:16:01.712 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:16:01.712 "subtype": "Discovery", 00:16:01.712 "listen_addresses": [], 00:16:01.712 "allow_any_host": true, 00:16:01.712 "hosts": [] 00:16:01.712 }, 00:16:01.712 { 00:16:01.712 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:16:01.712 "subtype": "NVMe", 00:16:01.712 "listen_addresses": [ 00:16:01.712 { 00:16:01.712 "trtype": "VFIOUSER", 00:16:01.712 "adrfam": "IPv4", 00:16:01.712 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:16:01.712 "trsvcid": "0" 00:16:01.712 } 00:16:01.712 ], 00:16:01.713 "allow_any_host": true, 00:16:01.713 "hosts": [], 00:16:01.713 "serial_number": "SPDK1", 00:16:01.713 "model_number": "SPDK bdev Controller", 00:16:01.713 "max_namespaces": 32, 00:16:01.713 "min_cntlid": 1, 00:16:01.713 "max_cntlid": 65519, 00:16:01.713 "namespaces": [ 00:16:01.713 { 00:16:01.713 "nsid": 1, 00:16:01.713 "bdev_name": "Malloc1", 00:16:01.713 "name": "Malloc1", 00:16:01.713 "nguid": "96B405269B204EC3961FED682E967251", 00:16:01.713 "uuid": "96b40526-9b20-4ec3-961f-ed682e967251" 00:16:01.713 }, 00:16:01.713 { 00:16:01.713 "nsid": 2, 00:16:01.713 "bdev_name": "Malloc3", 00:16:01.713 "name": "Malloc3", 00:16:01.713 "nguid": "7A771E4159474DA09551F7060D8371E1", 00:16:01.713 "uuid": "7a771e41-5947-4da0-9551-f7060d8371e1" 00:16:01.713 } 00:16:01.713 ] 00:16:01.713 }, 00:16:01.713 { 00:16:01.713 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:16:01.713 "subtype": "NVMe", 00:16:01.713 "listen_addresses": [ 00:16:01.713 { 00:16:01.713 "trtype": "VFIOUSER", 00:16:01.713 "adrfam": "IPv4", 00:16:01.713 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:16:01.713 "trsvcid": "0" 00:16:01.713 } 00:16:01.713 ], 00:16:01.713 "allow_any_host": true, 00:16:01.713 "hosts": [], 00:16:01.713 "serial_number": "SPDK2", 00:16:01.713 "model_number": "SPDK bdev Controller", 00:16:01.713 "max_namespaces": 32, 00:16:01.713 "min_cntlid": 1, 00:16:01.713 "max_cntlid": 65519, 00:16:01.713 "namespaces": [ 00:16:01.713 { 00:16:01.713 "nsid": 1, 00:16:01.713 "bdev_name": "Malloc2", 00:16:01.713 "name": "Malloc2", 00:16:01.713 "nguid": "67897639123E44E1BF3F80023698BF86", 00:16:01.713 "uuid": "67897639-123e-44e1-bf3f-80023698bf86" 00:16:01.713 } 00:16:01.713 ] 00:16:01.713 } 00:16:01.713 ] 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=3815056 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -n 2 -g -t /tmp/aer_touch_file 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1269 -- # local i=0 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1280 -- # return 0 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:16:01.713 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc4 00:16:01.974 Malloc4 00:16:01.974 [2024-11-26 19:12:14.412458] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:16:01.974 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc4 -n 2 00:16:01.974 [2024-11-26 19:12:14.590615] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:16:02.235 Asynchronous Event Request test 00:16:02.235 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:16:02.235 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:16:02.235 Registering asynchronous event callbacks... 00:16:02.235 Starting namespace attribute notice tests for all controllers... 00:16:02.235 /var/run/vfio-user/domain/vfio-user2/2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:16:02.235 aer_cb - Changed Namespace 00:16:02.235 Cleaning up... 00:16:02.235 [ 00:16:02.235 { 00:16:02.235 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:16:02.235 "subtype": "Discovery", 00:16:02.235 "listen_addresses": [], 00:16:02.235 "allow_any_host": true, 00:16:02.235 "hosts": [] 00:16:02.235 }, 00:16:02.235 { 00:16:02.235 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:16:02.235 "subtype": "NVMe", 00:16:02.235 "listen_addresses": [ 00:16:02.235 { 00:16:02.235 "trtype": "VFIOUSER", 00:16:02.235 "adrfam": "IPv4", 00:16:02.235 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:16:02.235 "trsvcid": "0" 00:16:02.235 } 00:16:02.235 ], 00:16:02.235 "allow_any_host": true, 00:16:02.235 "hosts": [], 00:16:02.235 "serial_number": "SPDK1", 00:16:02.235 "model_number": "SPDK bdev Controller", 00:16:02.235 "max_namespaces": 32, 00:16:02.235 "min_cntlid": 1, 00:16:02.235 "max_cntlid": 65519, 00:16:02.235 "namespaces": [ 00:16:02.235 { 00:16:02.235 "nsid": 1, 00:16:02.235 "bdev_name": "Malloc1", 00:16:02.235 "name": "Malloc1", 00:16:02.235 "nguid": "96B405269B204EC3961FED682E967251", 00:16:02.235 "uuid": "96b40526-9b20-4ec3-961f-ed682e967251" 00:16:02.235 }, 00:16:02.235 { 00:16:02.235 "nsid": 2, 00:16:02.235 "bdev_name": "Malloc3", 00:16:02.235 "name": "Malloc3", 00:16:02.235 "nguid": "7A771E4159474DA09551F7060D8371E1", 00:16:02.235 "uuid": "7a771e41-5947-4da0-9551-f7060d8371e1" 00:16:02.235 } 00:16:02.235 ] 00:16:02.235 }, 00:16:02.235 { 00:16:02.235 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:16:02.235 "subtype": "NVMe", 00:16:02.235 "listen_addresses": [ 00:16:02.235 { 00:16:02.235 "trtype": "VFIOUSER", 00:16:02.235 "adrfam": "IPv4", 00:16:02.235 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:16:02.235 "trsvcid": "0" 00:16:02.235 } 00:16:02.235 ], 00:16:02.235 "allow_any_host": true, 00:16:02.235 "hosts": [], 00:16:02.235 "serial_number": "SPDK2", 00:16:02.235 "model_number": "SPDK bdev Controller", 00:16:02.235 "max_namespaces": 32, 00:16:02.235 "min_cntlid": 1, 00:16:02.235 "max_cntlid": 65519, 00:16:02.235 "namespaces": [ 00:16:02.235 { 00:16:02.235 "nsid": 1, 00:16:02.235 "bdev_name": "Malloc2", 00:16:02.235 "name": "Malloc2", 00:16:02.235 "nguid": "67897639123E44E1BF3F80023698BF86", 00:16:02.235 "uuid": "67897639-123e-44e1-bf3f-80023698bf86" 00:16:02.235 }, 00:16:02.235 { 00:16:02.235 "nsid": 2, 00:16:02.235 "bdev_name": "Malloc4", 00:16:02.235 "name": "Malloc4", 00:16:02.235 "nguid": "A33338AC8C164013BA2E8722770EB8AF", 00:16:02.235 "uuid": "a33338ac-8c16-4013-ba2e-8722770eb8af" 00:16:02.235 } 00:16:02.235 ] 00:16:02.235 } 00:16:02.235 ] 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 3815056 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@105 -- # stop_nvmf_vfio_user 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 3805323 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # '[' -z 3805323 ']' 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@958 -- # kill -0 3805323 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # uname 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:02.235 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3805323 00:16:02.496 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:02.496 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:02.496 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3805323' 00:16:02.496 killing process with pid 3805323 00:16:02.496 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@973 -- # kill 3805323 00:16:02.496 19:12:14 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@978 -- # wait 3805323 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@108 -- # setup_nvmf_vfio_user --interrupt-mode '-M -I' 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args=--interrupt-mode 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local 'transport_args=-M -I' 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=3815391 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 3815391' 00:16:02.496 Process pid: 3815391 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:16:02.496 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' --interrupt-mode 00:16:02.497 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 3815391 00:16:02.497 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # '[' -z 3815391 ']' 00:16:02.497 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:02.497 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:02.497 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:02.497 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:02.497 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:02.497 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:16:02.497 [2024-11-26 19:12:15.096173] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:16:02.497 [2024-11-26 19:12:15.097119] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:16:02.497 [2024-11-26 19:12:15.097161] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:02.758 [2024-11-26 19:12:15.176660] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:16:02.758 [2024-11-26 19:12:15.211771] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:02.758 [2024-11-26 19:12:15.211809] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:02.758 [2024-11-26 19:12:15.211817] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:02.758 [2024-11-26 19:12:15.211823] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:02.758 [2024-11-26 19:12:15.211829] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:02.758 [2024-11-26 19:12:15.213336] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:02.758 [2024-11-26 19:12:15.213450] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:16:02.758 [2024-11-26 19:12:15.213603] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:02.758 [2024-11-26 19:12:15.213604] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:16:02.758 [2024-11-26 19:12:15.269883] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:16:02.758 [2024-11-26 19:12:15.270036] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:16:02.758 [2024-11-26 19:12:15.271063] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:16:02.758 [2024-11-26 19:12:15.271706] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:16:02.758 [2024-11-26 19:12:15.271784] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:16:03.329 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:03.329 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@868 -- # return 0 00:16:03.329 19:12:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:16:04.718 19:12:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER -M -I 00:16:04.718 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:16:04.718 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:16:04.718 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:16:04.718 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:16:04.718 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:16:04.718 Malloc1 00:16:04.718 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:16:04.979 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:16:05.241 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:16:05.502 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:16:05.502 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:16:05.502 19:12:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:16:05.502 Malloc2 00:16:05.502 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:16:05.762 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@109 -- # stop_nvmf_vfio_user 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 3815391 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # '[' -z 3815391 ']' 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@958 -- # kill -0 3815391 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # uname 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:06.023 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3815391 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3815391' 00:16:06.331 killing process with pid 3815391 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@973 -- # kill 3815391 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@978 -- # wait 3815391 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:16:06.331 00:16:06.331 real 0m51.478s 00:16:06.331 user 3m17.030s 00:16:06.331 sys 0m2.888s 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:06.331 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:16:06.332 ************************************ 00:16:06.332 END TEST nvmf_vfio_user 00:16:06.332 ************************************ 00:16:06.332 19:12:18 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@32 -- # run_test nvmf_vfio_user_nvme_compliance /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:16:06.332 19:12:18 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:06.332 19:12:18 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:06.332 19:12:18 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:16:06.332 ************************************ 00:16:06.332 START TEST nvmf_vfio_user_nvme_compliance 00:16:06.332 ************************************ 00:16:06.332 19:12:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:16:06.595 * Looking for test storage... 00:16:06.595 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1693 -- # lcov --version 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # IFS=.-: 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # read -ra ver1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # IFS=.-: 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # read -ra ver2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@338 -- # local 'op=<' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@340 -- # ver1_l=2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@341 -- # ver2_l=1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@344 -- # case "$op" in 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@345 -- # : 1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # decimal 1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # ver1[v]=1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # decimal 2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # ver2[v]=2 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # return 0 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:16:06.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:06.595 --rc genhtml_branch_coverage=1 00:16:06.595 --rc genhtml_function_coverage=1 00:16:06.595 --rc genhtml_legend=1 00:16:06.595 --rc geninfo_all_blocks=1 00:16:06.595 --rc geninfo_unexecuted_blocks=1 00:16:06.595 00:16:06.595 ' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:16:06.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:06.595 --rc genhtml_branch_coverage=1 00:16:06.595 --rc genhtml_function_coverage=1 00:16:06.595 --rc genhtml_legend=1 00:16:06.595 --rc geninfo_all_blocks=1 00:16:06.595 --rc geninfo_unexecuted_blocks=1 00:16:06.595 00:16:06.595 ' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:16:06.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:06.595 --rc genhtml_branch_coverage=1 00:16:06.595 --rc genhtml_function_coverage=1 00:16:06.595 --rc genhtml_legend=1 00:16:06.595 --rc geninfo_all_blocks=1 00:16:06.595 --rc geninfo_unexecuted_blocks=1 00:16:06.595 00:16:06.595 ' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:16:06.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:06.595 --rc genhtml_branch_coverage=1 00:16:06.595 --rc genhtml_function_coverage=1 00:16:06.595 --rc genhtml_legend=1 00:16:06.595 --rc geninfo_all_blocks=1 00:16:06.595 --rc geninfo_unexecuted_blocks=1 00:16:06.595 00:16:06.595 ' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # uname -s 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@15 -- # shopt -s extglob 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@5 -- # export PATH 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@51 -- # : 0 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:16:06.595 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@55 -- # have_pci_nics=0 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@11 -- # MALLOC_BDEV_SIZE=64 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # export TEST_TRANSPORT=VFIOUSER 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # TEST_TRANSPORT=VFIOUSER 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@16 -- # rm -rf /var/run/vfio-user 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@20 -- # nvmfpid=3816151 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@21 -- # echo 'Process pid: 3816151' 00:16:06.595 Process pid: 3816151 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@23 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@24 -- # waitforlisten 3816151 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@835 -- # '[' -z 3816151 ']' 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:06.595 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:06.595 19:12:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:16:06.855 [2024-11-26 19:12:19.219117] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:16:06.855 [2024-11-26 19:12:19.219175] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:06.855 [2024-11-26 19:12:19.300792] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:16:06.855 [2024-11-26 19:12:19.338267] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:06.855 [2024-11-26 19:12:19.338304] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:06.855 [2024-11-26 19:12:19.338312] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:06.855 [2024-11-26 19:12:19.338319] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:06.855 [2024-11-26 19:12:19.338325] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:06.855 [2024-11-26 19:12:19.339857] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:06.855 [2024-11-26 19:12:19.340003] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:16:06.855 [2024-11-26 19:12:19.340097] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:07.427 19:12:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:07.427 19:12:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@868 -- # return 0 00:16:07.427 19:12:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@26 -- # sleep 1 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@28 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@29 -- # traddr=/var/run/vfio-user 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@31 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@33 -- # mkdir -p /var/run/vfio-user 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@35 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:16:08.823 malloc0 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@36 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk -m 32 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@37 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@38 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.823 19:12:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/nvme_compliance -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user subnqn:nqn.2021-09.io.spdk:cnode0' 00:16:08.823 00:16:08.823 00:16:08.823 CUnit - A unit testing framework for C - Version 2.1-3 00:16:08.823 http://cunit.sourceforge.net/ 00:16:08.823 00:16:08.823 00:16:08.823 Suite: nvme_compliance 00:16:08.823 Test: admin_identify_ctrlr_verify_dptr ...[2024-11-26 19:12:21.300279] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:08.823 [2024-11-26 19:12:21.301670] vfio_user.c: 807:nvme_cmd_map_prps: *ERROR*: no PRP2, 3072 remaining 00:16:08.823 [2024-11-26 19:12:21.301683] vfio_user.c:5511:map_admin_cmd_req: *ERROR*: /var/run/vfio-user: map Admin Opc 6 failed 00:16:08.823 [2024-11-26 19:12:21.301688] vfio_user.c:5604:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x6 failed 00:16:08.823 [2024-11-26 19:12:21.304297] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:08.823 passed 00:16:08.823 Test: admin_identify_ctrlr_verify_fused ...[2024-11-26 19:12:21.398867] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:08.823 [2024-11-26 19:12:21.401890] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:08.823 passed 00:16:09.085 Test: admin_identify_ns ...[2024-11-26 19:12:21.499110] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.085 [2024-11-26 19:12:21.558878] ctrlr.c:2752:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 0 00:16:09.085 [2024-11-26 19:12:21.566876] ctrlr.c:2752:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 4294967295 00:16:09.085 [2024-11-26 19:12:21.587992] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.085 passed 00:16:09.085 Test: admin_get_features_mandatory_features ...[2024-11-26 19:12:21.682020] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.085 [2024-11-26 19:12:21.685038] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.347 passed 00:16:09.347 Test: admin_get_features_optional_features ...[2024-11-26 19:12:21.779590] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.347 [2024-11-26 19:12:21.782614] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.347 passed 00:16:09.347 Test: admin_set_features_number_of_queues ...[2024-11-26 19:12:21.875768] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.628 [2024-11-26 19:12:21.977969] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.628 passed 00:16:09.628 Test: admin_get_log_page_mandatory_logs ...[2024-11-26 19:12:22.071991] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.628 [2024-11-26 19:12:22.075011] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.628 passed 00:16:09.628 Test: admin_get_log_page_with_lpo ...[2024-11-26 19:12:22.168114] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.628 [2024-11-26 19:12:22.235875] ctrlr.c:2699:nvmf_ctrlr_get_log_page: *ERROR*: Get log page: offset (516) > len (512) 00:16:09.933 [2024-11-26 19:12:22.248919] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.933 passed 00:16:09.933 Test: fabric_property_get ...[2024-11-26 19:12:22.342970] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.933 [2024-11-26 19:12:22.344218] vfio_user.c:5604:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x7f failed 00:16:09.933 [2024-11-26 19:12:22.345987] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.933 passed 00:16:09.933 Test: admin_delete_io_sq_use_admin_qid ...[2024-11-26 19:12:22.441821] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:09.933 [2024-11-26 19:12:22.443069] vfio_user.c:2312:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:0 does not exist 00:16:09.933 [2024-11-26 19:12:22.444839] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:09.933 passed 00:16:09.933 Test: admin_delete_io_sq_delete_sq_twice ...[2024-11-26 19:12:22.537973] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:10.217 [2024-11-26 19:12:22.623873] vfio_user.c:2312:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:16:10.217 [2024-11-26 19:12:22.639870] vfio_user.c:2312:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:16:10.217 [2024-11-26 19:12:22.644953] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:10.217 passed 00:16:10.217 Test: admin_delete_io_cq_use_admin_qid ...[2024-11-26 19:12:22.734555] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:10.217 [2024-11-26 19:12:22.735800] vfio_user.c:2312:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O cqid:0 does not exist 00:16:10.217 [2024-11-26 19:12:22.737579] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:10.217 passed 00:16:10.496 Test: admin_delete_io_cq_delete_cq_first ...[2024-11-26 19:12:22.830697] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:10.496 [2024-11-26 19:12:22.906868] vfio_user.c:2322:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:16:10.496 [2024-11-26 19:12:22.930877] vfio_user.c:2312:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:16:10.496 [2024-11-26 19:12:22.935943] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:10.496 passed 00:16:10.496 Test: admin_create_io_cq_verify_iv_pc ...[2024-11-26 19:12:23.027561] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:10.496 [2024-11-26 19:12:23.028807] vfio_user.c:2161:handle_create_io_cq: *ERROR*: /var/run/vfio-user: IV is too big 00:16:10.496 [2024-11-26 19:12:23.028829] vfio_user.c:2155:handle_create_io_cq: *ERROR*: /var/run/vfio-user: non-PC CQ not supported 00:16:10.496 [2024-11-26 19:12:23.030576] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:10.496 passed 00:16:10.783 Test: admin_create_io_sq_verify_qsize_cqid ...[2024-11-26 19:12:23.124716] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:10.783 [2024-11-26 19:12:23.216874] vfio_user.c:2243:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 1 00:16:10.783 [2024-11-26 19:12:23.224871] vfio_user.c:2243:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 257 00:16:10.783 [2024-11-26 19:12:23.232871] vfio_user.c:2041:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:0 00:16:10.783 [2024-11-26 19:12:23.240876] vfio_user.c:2041:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:128 00:16:10.783 [2024-11-26 19:12:23.269957] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:10.783 passed 00:16:10.783 Test: admin_create_io_sq_verify_pc ...[2024-11-26 19:12:23.360591] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:10.783 [2024-11-26 19:12:23.375881] vfio_user.c:2054:handle_create_io_sq: *ERROR*: /var/run/vfio-user: non-PC SQ not supported 00:16:10.783 [2024-11-26 19:12:23.393734] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:11.045 passed 00:16:11.045 Test: admin_create_io_qp_max_qps ...[2024-11-26 19:12:23.487265] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:11.987 [2024-11-26 19:12:24.583877] nvme_ctrlr.c:5523:spdk_nvme_ctrlr_alloc_qid: *ERROR*: [/var/run/vfio-user, 0] No free I/O queue IDs 00:16:12.559 [2024-11-26 19:12:24.989823] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:12.559 passed 00:16:12.559 Test: admin_create_io_sq_shared_cq ...[2024-11-26 19:12:25.080953] vfio_user.c:2840:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:16:12.820 [2024-11-26 19:12:25.213881] vfio_user.c:2322:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:16:12.820 [2024-11-26 19:12:25.250924] vfio_user.c:2802:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:16:12.820 passed 00:16:12.820 00:16:12.820 Run Summary: Type Total Ran Passed Failed Inactive 00:16:12.820 suites 1 1 n/a 0 0 00:16:12.820 tests 18 18 18 0 0 00:16:12.820 asserts 360 360 360 0 n/a 00:16:12.820 00:16:12.820 Elapsed time = 1.655 seconds 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@42 -- # killprocess 3816151 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@954 -- # '[' -z 3816151 ']' 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@958 -- # kill -0 3816151 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@959 -- # uname 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3816151 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3816151' 00:16:12.820 killing process with pid 3816151 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@973 -- # kill 3816151 00:16:12.820 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@978 -- # wait 3816151 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@44 -- # rm -rf /var/run/vfio-user 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:16:13.116 00:16:13.116 real 0m6.590s 00:16:13.116 user 0m18.691s 00:16:13.116 sys 0m0.536s 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:16:13.116 ************************************ 00:16:13.116 END TEST nvmf_vfio_user_nvme_compliance 00:16:13.116 ************************************ 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@33 -- # run_test nvmf_vfio_user_fuzz /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:16:13.116 ************************************ 00:16:13.116 START TEST nvmf_vfio_user_fuzz 00:16:13.116 ************************************ 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:16:13.116 * Looking for test storage... 00:16:13.116 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1693 -- # lcov --version 00:16:13.116 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # IFS=.-: 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # read -ra ver1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # IFS=.-: 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # read -ra ver2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@338 -- # local 'op=<' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@340 -- # ver1_l=2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@341 -- # ver2_l=1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@344 -- # case "$op" in 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@345 -- # : 1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # decimal 1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # ver1[v]=1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # decimal 2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # ver2[v]=2 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # return 0 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:16:13.379 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:13.379 --rc genhtml_branch_coverage=1 00:16:13.379 --rc genhtml_function_coverage=1 00:16:13.379 --rc genhtml_legend=1 00:16:13.379 --rc geninfo_all_blocks=1 00:16:13.379 --rc geninfo_unexecuted_blocks=1 00:16:13.379 00:16:13.379 ' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:16:13.379 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:13.379 --rc genhtml_branch_coverage=1 00:16:13.379 --rc genhtml_function_coverage=1 00:16:13.379 --rc genhtml_legend=1 00:16:13.379 --rc geninfo_all_blocks=1 00:16:13.379 --rc geninfo_unexecuted_blocks=1 00:16:13.379 00:16:13.379 ' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:16:13.379 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:13.379 --rc genhtml_branch_coverage=1 00:16:13.379 --rc genhtml_function_coverage=1 00:16:13.379 --rc genhtml_legend=1 00:16:13.379 --rc geninfo_all_blocks=1 00:16:13.379 --rc geninfo_unexecuted_blocks=1 00:16:13.379 00:16:13.379 ' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:16:13.379 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:13.379 --rc genhtml_branch_coverage=1 00:16:13.379 --rc genhtml_function_coverage=1 00:16:13.379 --rc genhtml_legend=1 00:16:13.379 --rc geninfo_all_blocks=1 00:16:13.379 --rc geninfo_unexecuted_blocks=1 00:16:13.379 00:16:13.379 ' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # uname -s 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@15 -- # shopt -s extglob 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:13.379 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@5 -- # export PATH 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@51 -- # : 0 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:16:13.380 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@55 -- # have_pci_nics=0 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@12 -- # MALLOC_BDEV_SIZE=64 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@15 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@16 -- # traddr=/var/run/vfio-user 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@20 -- # rm -rf /var/run/vfio-user 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@24 -- # nvmfpid=3817579 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@25 -- # echo 'Process pid: 3817579' 00:16:13.380 Process pid: 3817579 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@27 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@28 -- # waitforlisten 3817579 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@835 -- # '[' -z 3817579 ']' 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:13.380 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:13.380 19:12:25 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:14.322 19:12:26 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:14.322 19:12:26 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@868 -- # return 0 00:16:14.322 19:12:26 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@30 -- # sleep 1 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@32 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@34 -- # mkdir -p /var/run/vfio-user 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:15.264 malloc0 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@39 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@41 -- # trid='trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' 00:16:15.264 19:12:27 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -t 30 -S 123456 -F 'trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' -N -a 00:16:47.394 Fuzzing completed. Shutting down the fuzz application 00:16:47.394 00:16:47.394 Dumping successful admin opcodes: 00:16:47.394 9, 10, 00:16:47.394 Dumping successful io opcodes: 00:16:47.394 0, 00:16:47.394 NS: 0x20000081ef00 I/O qp, Total commands completed: 1125571, total successful commands: 4432, random_seed: 429253888 00:16:47.394 NS: 0x20000081ef00 admin qp, Total commands completed: 142784, total successful commands: 32, random_seed: 1021563136 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@44 -- # rpc_cmd nvmf_delete_subsystem nqn.2021-09.io.spdk:cnode0 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@46 -- # killprocess 3817579 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@954 -- # '[' -z 3817579 ']' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@958 -- # kill -0 3817579 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@959 -- # uname 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3817579 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3817579' 00:16:47.394 killing process with pid 3817579 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@973 -- # kill 3817579 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@978 -- # wait 3817579 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@48 -- # rm -rf /var/run/vfio-user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_log.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_tgt_output.txt 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@50 -- # trap - SIGINT SIGTERM EXIT 00:16:47.394 00:16:47.394 real 0m33.799s 00:16:47.394 user 0m37.671s 00:16:47.394 sys 0m26.966s 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:16:47.394 ************************************ 00:16:47.394 END TEST nvmf_vfio_user_fuzz 00:16:47.394 ************************************ 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@37 -- # run_test nvmf_auth_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:16:47.394 ************************************ 00:16:47.394 START TEST nvmf_auth_target 00:16:47.394 ************************************ 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:16:47.394 * Looking for test storage... 00:16:47.394 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1693 -- # lcov --version 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # IFS=.-: 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # read -ra ver1 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # IFS=.-: 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # read -ra ver2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@338 -- # local 'op=<' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@340 -- # ver1_l=2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@341 -- # ver2_l=1 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@344 -- # case "$op" in 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@345 -- # : 1 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # decimal 1 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=1 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 1 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # ver1[v]=1 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # decimal 2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # ver2[v]=2 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # return 0 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:16:47.394 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.394 --rc genhtml_branch_coverage=1 00:16:47.394 --rc genhtml_function_coverage=1 00:16:47.394 --rc genhtml_legend=1 00:16:47.394 --rc geninfo_all_blocks=1 00:16:47.394 --rc geninfo_unexecuted_blocks=1 00:16:47.394 00:16:47.394 ' 00:16:47.394 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:16:47.394 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.395 --rc genhtml_branch_coverage=1 00:16:47.395 --rc genhtml_function_coverage=1 00:16:47.395 --rc genhtml_legend=1 00:16:47.395 --rc geninfo_all_blocks=1 00:16:47.395 --rc geninfo_unexecuted_blocks=1 00:16:47.395 00:16:47.395 ' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:16:47.395 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.395 --rc genhtml_branch_coverage=1 00:16:47.395 --rc genhtml_function_coverage=1 00:16:47.395 --rc genhtml_legend=1 00:16:47.395 --rc geninfo_all_blocks=1 00:16:47.395 --rc geninfo_unexecuted_blocks=1 00:16:47.395 00:16:47.395 ' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:16:47.395 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.395 --rc genhtml_branch_coverage=1 00:16:47.395 --rc genhtml_function_coverage=1 00:16:47.395 --rc genhtml_legend=1 00:16:47.395 --rc geninfo_all_blocks=1 00:16:47.395 --rc geninfo_unexecuted_blocks=1 00:16:47.395 00:16:47.395 ' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # uname -s 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@15 -- # shopt -s extglob 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@5 -- # export PATH 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@51 -- # : 0 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:16:47.395 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@14 -- # dhgroups=("null" "ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@15 -- # subnqn=nqn.2024-03.io.spdk:cnode0 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@16 -- # hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@17 -- # hostsock=/var/tmp/host.sock 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # keys=() 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # ckeys=() 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@86 -- # nvmftestinit 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@309 -- # xtrace_disable 00:16:47.395 19:12:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # pci_devs=() 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # net_devs=() 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # e810=() 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # local -ga e810 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # x722=() 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # local -ga x722 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # mlx=() 00:16:55.533 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # local -ga mlx 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:16:55.534 Found 0000:31:00.0 (0x8086 - 0x159b) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:16:55.534 Found 0000:31:00.1 (0x8086 - 0x159b) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:16:55.534 Found net devices under 0000:31:00.0: cvl_0_0 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:16:55.534 Found net devices under 0000:31:00.1: cvl_0_1 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # is_hw=yes 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:16:55.534 19:13:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:16:55.534 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:16:55.534 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:16:55.534 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:16:55.534 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:16:55.534 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:16:55.534 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.688 ms 00:16:55.534 00:16:55.534 --- 10.0.0.2 ping statistics --- 00:16:55.534 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:55.534 rtt min/avg/max/mdev = 0.688/0.688/0.688/0.000 ms 00:16:55.534 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:16:55.534 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:16:55.534 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.320 ms 00:16:55.534 00:16:55.534 --- 10.0.0.1 ping statistics --- 00:16:55.534 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:55.534 rtt min/avg/max/mdev = 0.320/0.320/0.320/0.000 ms 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@450 -- # return 0 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@87 -- # nvmfappstart -L nvmf_auth 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # nvmfpid=3828270 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # waitforlisten 3828270 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvmf_auth 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 3828270 ']' 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:55.535 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:56.476 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:56.476 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:16:56.476 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:16:56.476 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:56.476 19:13:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@89 -- # hostpid=3828591 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@91 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 2 -r /var/tmp/host.sock -L nvme_auth 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key null 48 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=null 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=48 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=60785d69afdc306a59c1571fa44b777a50c508eaeeb6d132 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.uaB 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 60785d69afdc306a59c1571fa44b777a50c508eaeeb6d132 0 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 60785d69afdc306a59c1571fa44b777a50c508eaeeb6d132 0 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=60785d69afdc306a59c1571fa44b777a50c508eaeeb6d132 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=0 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.uaB 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.uaB 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # keys[0]=/tmp/spdk.key-null.uaB 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key sha512 64 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha512 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=64 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:16:56.476 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=90a24cb70c0ab400cdd4ac5ac2fda9c34912f707298679dc57a9d271f92a3104 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.tl0 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 90a24cb70c0ab400cdd4ac5ac2fda9c34912f707298679dc57a9d271f92a3104 3 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 90a24cb70c0ab400cdd4ac5ac2fda9c34912f707298679dc57a9d271f92a3104 3 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=90a24cb70c0ab400cdd4ac5ac2fda9c34912f707298679dc57a9d271f92a3104 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=3 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.tl0 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.tl0 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # ckeys[0]=/tmp/spdk.key-sha512.tl0 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha256 32 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha256 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=32 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=3ce7ac4b6b0c72fcee9ddd93d7b442f8 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.SXT 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 3ce7ac4b6b0c72fcee9ddd93d7b442f8 1 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 3ce7ac4b6b0c72fcee9ddd93d7b442f8 1 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=3ce7ac4b6b0c72fcee9ddd93d7b442f8 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=1 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.SXT 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.SXT 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # keys[1]=/tmp/spdk.key-sha256.SXT 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha384 48 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha384 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=48 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=27ff4f79ee1d5711e832ce3642d2e4a0aa4b6c1914a8c402 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.PND 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 27ff4f79ee1d5711e832ce3642d2e4a0aa4b6c1914a8c402 2 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 27ff4f79ee1d5711e832ce3642d2e4a0aa4b6c1914a8c402 2 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=27ff4f79ee1d5711e832ce3642d2e4a0aa4b6c1914a8c402 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=2 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.PND 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.PND 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # ckeys[1]=/tmp/spdk.key-sha384.PND 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha384 48 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha384 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=48 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=a59d1aa408004349b42a2311556c21c5e3d0a555d45c01f4 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.0uM 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key a59d1aa408004349b42a2311556c21c5e3d0a555d45c01f4 2 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 a59d1aa408004349b42a2311556c21c5e3d0a555d45c01f4 2 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=a59d1aa408004349b42a2311556c21c5e3d0a555d45c01f4 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=2 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.0uM 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.0uM 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # keys[2]=/tmp/spdk.key-sha384.0uM 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha256 32 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha256 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=32 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=3ec098b97d1bc067b3af51e2f6685bda 00:16:56.737 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:16:56.997 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.D9C 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 3ec098b97d1bc067b3af51e2f6685bda 1 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 3ec098b97d1bc067b3af51e2f6685bda 1 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=3ec098b97d1bc067b3af51e2f6685bda 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=1 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.D9C 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.D9C 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # ckeys[2]=/tmp/spdk.key-sha256.D9C 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # gen_dhchap_key sha512 64 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha512 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=64 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=0087e1754afd58d7f4c61c1496e384a0e48fc4c2b44d3791887cd19b7ec572db 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.gQD 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 0087e1754afd58d7f4c61c1496e384a0e48fc4c2b44d3791887cd19b7ec572db 3 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 0087e1754afd58d7f4c61c1496e384a0e48fc4c2b44d3791887cd19b7ec572db 3 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=0087e1754afd58d7f4c61c1496e384a0e48fc4c2b44d3791887cd19b7ec572db 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=3 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.gQD 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.gQD 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # keys[3]=/tmp/spdk.key-sha512.gQD 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # ckeys[3]= 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@99 -- # waitforlisten 3828270 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 3828270 ']' 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:56.998 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:56.998 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@100 -- # waitforlisten 3828591 /var/tmp/host.sock 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 3828591 ']' 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/host.sock 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:16:57.258 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@101 -- # rpc_cmd 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.uaB 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key0 /tmp/spdk.key-null.uaB 00:16:57.258 19:13:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key0 /tmp/spdk.key-null.uaB 00:16:57.519 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha512.tl0 ]] 00:16:57.519 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.tl0 00:16:57.519 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.519 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.519 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.519 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey0 /tmp/spdk.key-sha512.tl0 00:16:57.519 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey0 /tmp/spdk.key-sha512.tl0 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.SXT 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key1 /tmp/spdk.key-sha256.SXT 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key1 /tmp/spdk.key-sha256.SXT 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha384.PND ]] 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.PND 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey1 /tmp/spdk.key-sha384.PND 00:16:57.779 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey1 /tmp/spdk.key-sha384.PND 00:16:58.039 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:16:58.039 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.0uM 00:16:58.039 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.039 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.039 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.039 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key2 /tmp/spdk.key-sha384.0uM 00:16:58.039 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key2 /tmp/spdk.key-sha384.0uM 00:16:58.299 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha256.D9C ]] 00:16:58.299 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.D9C 00:16:58.299 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.299 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.299 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.299 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey2 /tmp/spdk.key-sha256.D9C 00:16:58.299 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey2 /tmp/spdk.key-sha256.D9C 00:16:58.560 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:16:58.560 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.gQD 00:16:58.560 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.560 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.560 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.560 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key3 /tmp/spdk.key-sha512.gQD 00:16:58.560 19:13:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key3 /tmp/spdk.key-sha512.gQD 00:16:58.560 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n '' ]] 00:16:58.560 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:16:58.560 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:58.560 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:58.560 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:58.560 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 0 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.820 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:58.821 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:58.821 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:59.080 00:16:59.080 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:59.080 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:59.080 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:59.341 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:59.341 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:59.341 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:59.341 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:59.342 { 00:16:59.342 "cntlid": 1, 00:16:59.342 "qid": 0, 00:16:59.342 "state": "enabled", 00:16:59.342 "thread": "nvmf_tgt_poll_group_000", 00:16:59.342 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:16:59.342 "listen_address": { 00:16:59.342 "trtype": "TCP", 00:16:59.342 "adrfam": "IPv4", 00:16:59.342 "traddr": "10.0.0.2", 00:16:59.342 "trsvcid": "4420" 00:16:59.342 }, 00:16:59.342 "peer_address": { 00:16:59.342 "trtype": "TCP", 00:16:59.342 "adrfam": "IPv4", 00:16:59.342 "traddr": "10.0.0.1", 00:16:59.342 "trsvcid": "59712" 00:16:59.342 }, 00:16:59.342 "auth": { 00:16:59.342 "state": "completed", 00:16:59.342 "digest": "sha256", 00:16:59.342 "dhgroup": "null" 00:16:59.342 } 00:16:59.342 } 00:16:59.342 ]' 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:59.342 19:13:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:59.602 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:16:59.602 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:00.171 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:00.431 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 1 00:17:00.431 19:13:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:00.431 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:00.692 00:17:00.692 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:00.692 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:00.692 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:00.953 { 00:17:00.953 "cntlid": 3, 00:17:00.953 "qid": 0, 00:17:00.953 "state": "enabled", 00:17:00.953 "thread": "nvmf_tgt_poll_group_000", 00:17:00.953 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:00.953 "listen_address": { 00:17:00.953 "trtype": "TCP", 00:17:00.953 "adrfam": "IPv4", 00:17:00.953 "traddr": "10.0.0.2", 00:17:00.953 "trsvcid": "4420" 00:17:00.953 }, 00:17:00.953 "peer_address": { 00:17:00.953 "trtype": "TCP", 00:17:00.953 "adrfam": "IPv4", 00:17:00.953 "traddr": "10.0.0.1", 00:17:00.953 "trsvcid": "60682" 00:17:00.953 }, 00:17:00.953 "auth": { 00:17:00.953 "state": "completed", 00:17:00.953 "digest": "sha256", 00:17:00.953 "dhgroup": "null" 00:17:00.953 } 00:17:00.953 } 00:17:00.953 ]' 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:00.953 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:01.215 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:01.215 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:01.215 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:01.215 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:01.215 19:13:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:02.159 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 2 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.159 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.418 00:17:02.419 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:02.419 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:02.419 19:13:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:02.679 { 00:17:02.679 "cntlid": 5, 00:17:02.679 "qid": 0, 00:17:02.679 "state": "enabled", 00:17:02.679 "thread": "nvmf_tgt_poll_group_000", 00:17:02.679 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:02.679 "listen_address": { 00:17:02.679 "trtype": "TCP", 00:17:02.679 "adrfam": "IPv4", 00:17:02.679 "traddr": "10.0.0.2", 00:17:02.679 "trsvcid": "4420" 00:17:02.679 }, 00:17:02.679 "peer_address": { 00:17:02.679 "trtype": "TCP", 00:17:02.679 "adrfam": "IPv4", 00:17:02.679 "traddr": "10.0.0.1", 00:17:02.679 "trsvcid": "60712" 00:17:02.679 }, 00:17:02.679 "auth": { 00:17:02.679 "state": "completed", 00:17:02.679 "digest": "sha256", 00:17:02.679 "dhgroup": "null" 00:17:02.679 } 00:17:02.679 } 00:17:02.679 ]' 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:02.679 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:02.939 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:02.939 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:02.939 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:02.939 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:02.939 19:13:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:03.881 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 3 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:03.881 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:04.142 00:17:04.142 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:04.142 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:04.142 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:04.404 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:04.404 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:04.404 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.404 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:04.404 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:04.405 { 00:17:04.405 "cntlid": 7, 00:17:04.405 "qid": 0, 00:17:04.405 "state": "enabled", 00:17:04.405 "thread": "nvmf_tgt_poll_group_000", 00:17:04.405 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:04.405 "listen_address": { 00:17:04.405 "trtype": "TCP", 00:17:04.405 "adrfam": "IPv4", 00:17:04.405 "traddr": "10.0.0.2", 00:17:04.405 "trsvcid": "4420" 00:17:04.405 }, 00:17:04.405 "peer_address": { 00:17:04.405 "trtype": "TCP", 00:17:04.405 "adrfam": "IPv4", 00:17:04.405 "traddr": "10.0.0.1", 00:17:04.405 "trsvcid": "60734" 00:17:04.405 }, 00:17:04.405 "auth": { 00:17:04.405 "state": "completed", 00:17:04.405 "digest": "sha256", 00:17:04.405 "dhgroup": "null" 00:17:04.405 } 00:17:04.405 } 00:17:04.405 ]' 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:04.405 19:13:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:04.665 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:04.665 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:05.607 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:05.607 19:13:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:05.607 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 0 00:17:05.607 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:05.607 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:05.607 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:05.607 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:05.607 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:05.608 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:05.608 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.608 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:05.608 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.608 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:05.608 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:05.608 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:05.867 00:17:05.867 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:05.867 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:05.867 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:06.127 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:06.127 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:06.127 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:06.127 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:06.127 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:06.127 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:06.127 { 00:17:06.127 "cntlid": 9, 00:17:06.127 "qid": 0, 00:17:06.127 "state": "enabled", 00:17:06.127 "thread": "nvmf_tgt_poll_group_000", 00:17:06.128 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:06.128 "listen_address": { 00:17:06.128 "trtype": "TCP", 00:17:06.128 "adrfam": "IPv4", 00:17:06.128 "traddr": "10.0.0.2", 00:17:06.128 "trsvcid": "4420" 00:17:06.128 }, 00:17:06.128 "peer_address": { 00:17:06.128 "trtype": "TCP", 00:17:06.128 "adrfam": "IPv4", 00:17:06.128 "traddr": "10.0.0.1", 00:17:06.128 "trsvcid": "60760" 00:17:06.128 }, 00:17:06.128 "auth": { 00:17:06.128 "state": "completed", 00:17:06.128 "digest": "sha256", 00:17:06.128 "dhgroup": "ffdhe2048" 00:17:06.128 } 00:17:06.128 } 00:17:06.128 ]' 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:06.128 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:06.389 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:06.389 19:13:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:06.961 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:07.222 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 1 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:07.222 19:13:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:07.483 00:17:07.483 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:07.483 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:07.483 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:07.744 { 00:17:07.744 "cntlid": 11, 00:17:07.744 "qid": 0, 00:17:07.744 "state": "enabled", 00:17:07.744 "thread": "nvmf_tgt_poll_group_000", 00:17:07.744 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:07.744 "listen_address": { 00:17:07.744 "trtype": "TCP", 00:17:07.744 "adrfam": "IPv4", 00:17:07.744 "traddr": "10.0.0.2", 00:17:07.744 "trsvcid": "4420" 00:17:07.744 }, 00:17:07.744 "peer_address": { 00:17:07.744 "trtype": "TCP", 00:17:07.744 "adrfam": "IPv4", 00:17:07.744 "traddr": "10.0.0.1", 00:17:07.744 "trsvcid": "60782" 00:17:07.744 }, 00:17:07.744 "auth": { 00:17:07.744 "state": "completed", 00:17:07.744 "digest": "sha256", 00:17:07.744 "dhgroup": "ffdhe2048" 00:17:07.744 } 00:17:07.744 } 00:17:07.744 ]' 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:07.744 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:08.005 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:08.005 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:08.005 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:08.005 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:08.005 19:13:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:08.947 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 2 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:08.947 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:09.208 00:17:09.208 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:09.208 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:09.208 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:09.469 { 00:17:09.469 "cntlid": 13, 00:17:09.469 "qid": 0, 00:17:09.469 "state": "enabled", 00:17:09.469 "thread": "nvmf_tgt_poll_group_000", 00:17:09.469 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:09.469 "listen_address": { 00:17:09.469 "trtype": "TCP", 00:17:09.469 "adrfam": "IPv4", 00:17:09.469 "traddr": "10.0.0.2", 00:17:09.469 "trsvcid": "4420" 00:17:09.469 }, 00:17:09.469 "peer_address": { 00:17:09.469 "trtype": "TCP", 00:17:09.469 "adrfam": "IPv4", 00:17:09.469 "traddr": "10.0.0.1", 00:17:09.469 "trsvcid": "60800" 00:17:09.469 }, 00:17:09.469 "auth": { 00:17:09.469 "state": "completed", 00:17:09.469 "digest": "sha256", 00:17:09.469 "dhgroup": "ffdhe2048" 00:17:09.469 } 00:17:09.469 } 00:17:09.469 ]' 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:09.469 19:13:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:09.469 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:09.469 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:09.469 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:09.730 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:09.730 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:10.302 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:10.563 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:10.563 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:10.563 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:10.563 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:10.563 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:10.563 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:10.563 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:10.563 19:13:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:17:10.563 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 3 00:17:10.563 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:10.563 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:10.563 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:10.563 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:10.563 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:10.564 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:10.564 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:10.564 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:10.564 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:10.564 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:10.564 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:10.564 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:10.824 00:17:10.824 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:10.824 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:10.824 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:11.085 { 00:17:11.085 "cntlid": 15, 00:17:11.085 "qid": 0, 00:17:11.085 "state": "enabled", 00:17:11.085 "thread": "nvmf_tgt_poll_group_000", 00:17:11.085 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:11.085 "listen_address": { 00:17:11.085 "trtype": "TCP", 00:17:11.085 "adrfam": "IPv4", 00:17:11.085 "traddr": "10.0.0.2", 00:17:11.085 "trsvcid": "4420" 00:17:11.085 }, 00:17:11.085 "peer_address": { 00:17:11.085 "trtype": "TCP", 00:17:11.085 "adrfam": "IPv4", 00:17:11.085 "traddr": "10.0.0.1", 00:17:11.085 "trsvcid": "57356" 00:17:11.085 }, 00:17:11.085 "auth": { 00:17:11.085 "state": "completed", 00:17:11.085 "digest": "sha256", 00:17:11.085 "dhgroup": "ffdhe2048" 00:17:11.085 } 00:17:11.085 } 00:17:11.085 ]' 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:11.085 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:11.346 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:11.346 19:13:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:12.387 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 0 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:12.387 19:13:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:12.646 00:17:12.646 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:12.646 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:12.646 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:12.646 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:12.646 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:12.646 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.646 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:12.962 { 00:17:12.962 "cntlid": 17, 00:17:12.962 "qid": 0, 00:17:12.962 "state": "enabled", 00:17:12.962 "thread": "nvmf_tgt_poll_group_000", 00:17:12.962 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:12.962 "listen_address": { 00:17:12.962 "trtype": "TCP", 00:17:12.962 "adrfam": "IPv4", 00:17:12.962 "traddr": "10.0.0.2", 00:17:12.962 "trsvcid": "4420" 00:17:12.962 }, 00:17:12.962 "peer_address": { 00:17:12.962 "trtype": "TCP", 00:17:12.962 "adrfam": "IPv4", 00:17:12.962 "traddr": "10.0.0.1", 00:17:12.962 "trsvcid": "57388" 00:17:12.962 }, 00:17:12.962 "auth": { 00:17:12.962 "state": "completed", 00:17:12.962 "digest": "sha256", 00:17:12.962 "dhgroup": "ffdhe3072" 00:17:12.962 } 00:17:12.962 } 00:17:12.962 ]' 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:12.962 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:13.222 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:13.222 19:13:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:13.792 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:13.792 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 1 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:14.051 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:14.310 00:17:14.310 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:14.310 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:14.310 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:14.570 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:14.571 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:14.571 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.571 19:13:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:14.571 { 00:17:14.571 "cntlid": 19, 00:17:14.571 "qid": 0, 00:17:14.571 "state": "enabled", 00:17:14.571 "thread": "nvmf_tgt_poll_group_000", 00:17:14.571 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:14.571 "listen_address": { 00:17:14.571 "trtype": "TCP", 00:17:14.571 "adrfam": "IPv4", 00:17:14.571 "traddr": "10.0.0.2", 00:17:14.571 "trsvcid": "4420" 00:17:14.571 }, 00:17:14.571 "peer_address": { 00:17:14.571 "trtype": "TCP", 00:17:14.571 "adrfam": "IPv4", 00:17:14.571 "traddr": "10.0.0.1", 00:17:14.571 "trsvcid": "57408" 00:17:14.571 }, 00:17:14.571 "auth": { 00:17:14.571 "state": "completed", 00:17:14.571 "digest": "sha256", 00:17:14.571 "dhgroup": "ffdhe3072" 00:17:14.571 } 00:17:14.571 } 00:17:14.571 ]' 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:14.571 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:14.832 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:14.832 19:13:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:15.771 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 2 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:15.771 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.031 00:17:16.031 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:16.031 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:16.031 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:16.290 { 00:17:16.290 "cntlid": 21, 00:17:16.290 "qid": 0, 00:17:16.290 "state": "enabled", 00:17:16.290 "thread": "nvmf_tgt_poll_group_000", 00:17:16.290 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:16.290 "listen_address": { 00:17:16.290 "trtype": "TCP", 00:17:16.290 "adrfam": "IPv4", 00:17:16.290 "traddr": "10.0.0.2", 00:17:16.290 "trsvcid": "4420" 00:17:16.290 }, 00:17:16.290 "peer_address": { 00:17:16.290 "trtype": "TCP", 00:17:16.290 "adrfam": "IPv4", 00:17:16.290 "traddr": "10.0.0.1", 00:17:16.290 "trsvcid": "57434" 00:17:16.290 }, 00:17:16.290 "auth": { 00:17:16.290 "state": "completed", 00:17:16.290 "digest": "sha256", 00:17:16.290 "dhgroup": "ffdhe3072" 00:17:16.290 } 00:17:16.290 } 00:17:16.290 ]' 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:16.290 19:13:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:16.550 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:16.550 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:17.492 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 3 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:17.492 19:13:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:17.492 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:17.492 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.492 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:17.492 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.492 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:17.492 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:17.492 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:17.753 00:17:17.753 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:17.753 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:17.753 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:18.014 { 00:17:18.014 "cntlid": 23, 00:17:18.014 "qid": 0, 00:17:18.014 "state": "enabled", 00:17:18.014 "thread": "nvmf_tgt_poll_group_000", 00:17:18.014 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:18.014 "listen_address": { 00:17:18.014 "trtype": "TCP", 00:17:18.014 "adrfam": "IPv4", 00:17:18.014 "traddr": "10.0.0.2", 00:17:18.014 "trsvcid": "4420" 00:17:18.014 }, 00:17:18.014 "peer_address": { 00:17:18.014 "trtype": "TCP", 00:17:18.014 "adrfam": "IPv4", 00:17:18.014 "traddr": "10.0.0.1", 00:17:18.014 "trsvcid": "57464" 00:17:18.014 }, 00:17:18.014 "auth": { 00:17:18.014 "state": "completed", 00:17:18.014 "digest": "sha256", 00:17:18.014 "dhgroup": "ffdhe3072" 00:17:18.014 } 00:17:18.014 } 00:17:18.014 ]' 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:18.014 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:18.275 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:18.275 19:13:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:19.215 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 0 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:19.215 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:19.216 19:13:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:19.475 00:17:19.475 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:19.475 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:19.475 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:19.736 { 00:17:19.736 "cntlid": 25, 00:17:19.736 "qid": 0, 00:17:19.736 "state": "enabled", 00:17:19.736 "thread": "nvmf_tgt_poll_group_000", 00:17:19.736 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:19.736 "listen_address": { 00:17:19.736 "trtype": "TCP", 00:17:19.736 "adrfam": "IPv4", 00:17:19.736 "traddr": "10.0.0.2", 00:17:19.736 "trsvcid": "4420" 00:17:19.736 }, 00:17:19.736 "peer_address": { 00:17:19.736 "trtype": "TCP", 00:17:19.736 "adrfam": "IPv4", 00:17:19.736 "traddr": "10.0.0.1", 00:17:19.736 "trsvcid": "52170" 00:17:19.736 }, 00:17:19.736 "auth": { 00:17:19.736 "state": "completed", 00:17:19.736 "digest": "sha256", 00:17:19.736 "dhgroup": "ffdhe4096" 00:17:19.736 } 00:17:19.736 } 00:17:19.736 ]' 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:19.736 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:19.997 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:19.997 19:13:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:20.936 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:20.936 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:20.936 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:20.936 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.936 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:20.936 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.936 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 1 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:20.937 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:21.197 00:17:21.197 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:21.197 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:21.197 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:21.457 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:21.457 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:21.457 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.457 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:21.457 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.457 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:21.457 { 00:17:21.457 "cntlid": 27, 00:17:21.457 "qid": 0, 00:17:21.457 "state": "enabled", 00:17:21.457 "thread": "nvmf_tgt_poll_group_000", 00:17:21.457 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:21.457 "listen_address": { 00:17:21.457 "trtype": "TCP", 00:17:21.457 "adrfam": "IPv4", 00:17:21.457 "traddr": "10.0.0.2", 00:17:21.457 "trsvcid": "4420" 00:17:21.457 }, 00:17:21.457 "peer_address": { 00:17:21.457 "trtype": "TCP", 00:17:21.457 "adrfam": "IPv4", 00:17:21.457 "traddr": "10.0.0.1", 00:17:21.457 "trsvcid": "52198" 00:17:21.457 }, 00:17:21.457 "auth": { 00:17:21.457 "state": "completed", 00:17:21.457 "digest": "sha256", 00:17:21.457 "dhgroup": "ffdhe4096" 00:17:21.457 } 00:17:21.457 } 00:17:21.457 ]' 00:17:21.457 19:13:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:21.457 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:21.457 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:21.457 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:21.457 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:21.717 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:21.717 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:21.717 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:21.717 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:21.717 19:13:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:22.659 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 2 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:22.659 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:22.660 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.660 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:22.660 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.660 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:22.660 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:22.660 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:22.920 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:23.179 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:23.179 { 00:17:23.179 "cntlid": 29, 00:17:23.179 "qid": 0, 00:17:23.179 "state": "enabled", 00:17:23.179 "thread": "nvmf_tgt_poll_group_000", 00:17:23.179 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:23.179 "listen_address": { 00:17:23.179 "trtype": "TCP", 00:17:23.179 "adrfam": "IPv4", 00:17:23.179 "traddr": "10.0.0.2", 00:17:23.179 "trsvcid": "4420" 00:17:23.179 }, 00:17:23.179 "peer_address": { 00:17:23.179 "trtype": "TCP", 00:17:23.179 "adrfam": "IPv4", 00:17:23.179 "traddr": "10.0.0.1", 00:17:23.179 "trsvcid": "52232" 00:17:23.179 }, 00:17:23.179 "auth": { 00:17:23.179 "state": "completed", 00:17:23.179 "digest": "sha256", 00:17:23.179 "dhgroup": "ffdhe4096" 00:17:23.179 } 00:17:23.179 } 00:17:23.179 ]' 00:17:23.180 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:23.180 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:23.180 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:23.440 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:23.440 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:23.440 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:23.440 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:23.440 19:13:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:23.440 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:23.440 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:24.381 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:24.381 19:13:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 3 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:24.641 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:24.901 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.901 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:24.901 { 00:17:24.901 "cntlid": 31, 00:17:24.901 "qid": 0, 00:17:24.901 "state": "enabled", 00:17:24.901 "thread": "nvmf_tgt_poll_group_000", 00:17:24.901 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:24.901 "listen_address": { 00:17:24.901 "trtype": "TCP", 00:17:24.901 "adrfam": "IPv4", 00:17:24.901 "traddr": "10.0.0.2", 00:17:24.901 "trsvcid": "4420" 00:17:24.901 }, 00:17:24.901 "peer_address": { 00:17:24.901 "trtype": "TCP", 00:17:24.901 "adrfam": "IPv4", 00:17:24.901 "traddr": "10.0.0.1", 00:17:24.901 "trsvcid": "52252" 00:17:24.901 }, 00:17:24.901 "auth": { 00:17:24.902 "state": "completed", 00:17:24.902 "digest": "sha256", 00:17:24.902 "dhgroup": "ffdhe4096" 00:17:24.902 } 00:17:24.902 } 00:17:24.902 ]' 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:25.162 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:25.423 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:25.423 19:13:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:25.993 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:25.993 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 0 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.253 19:13:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.513 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:26.771 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:26.771 { 00:17:26.771 "cntlid": 33, 00:17:26.771 "qid": 0, 00:17:26.771 "state": "enabled", 00:17:26.771 "thread": "nvmf_tgt_poll_group_000", 00:17:26.771 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:26.771 "listen_address": { 00:17:26.771 "trtype": "TCP", 00:17:26.771 "adrfam": "IPv4", 00:17:26.771 "traddr": "10.0.0.2", 00:17:26.771 "trsvcid": "4420" 00:17:26.771 }, 00:17:26.771 "peer_address": { 00:17:26.771 "trtype": "TCP", 00:17:26.771 "adrfam": "IPv4", 00:17:26.771 "traddr": "10.0.0.1", 00:17:26.771 "trsvcid": "52284" 00:17:26.771 }, 00:17:26.771 "auth": { 00:17:26.771 "state": "completed", 00:17:26.771 "digest": "sha256", 00:17:26.771 "dhgroup": "ffdhe6144" 00:17:26.771 } 00:17:26.771 } 00:17:26.771 ]' 00:17:26.772 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:26.772 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:27.031 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:27.031 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:27.031 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:27.031 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:27.032 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:27.032 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:27.291 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:27.292 19:13:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:27.862 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:27.862 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:28.122 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 1 00:17:28.122 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:28.122 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:28.123 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:28.382 00:17:28.382 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:28.383 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:28.383 19:13:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:28.643 { 00:17:28.643 "cntlid": 35, 00:17:28.643 "qid": 0, 00:17:28.643 "state": "enabled", 00:17:28.643 "thread": "nvmf_tgt_poll_group_000", 00:17:28.643 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:28.643 "listen_address": { 00:17:28.643 "trtype": "TCP", 00:17:28.643 "adrfam": "IPv4", 00:17:28.643 "traddr": "10.0.0.2", 00:17:28.643 "trsvcid": "4420" 00:17:28.643 }, 00:17:28.643 "peer_address": { 00:17:28.643 "trtype": "TCP", 00:17:28.643 "adrfam": "IPv4", 00:17:28.643 "traddr": "10.0.0.1", 00:17:28.643 "trsvcid": "52304" 00:17:28.643 }, 00:17:28.643 "auth": { 00:17:28.643 "state": "completed", 00:17:28.643 "digest": "sha256", 00:17:28.643 "dhgroup": "ffdhe6144" 00:17:28.643 } 00:17:28.643 } 00:17:28.643 ]' 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:28.643 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:28.903 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:28.903 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:28.903 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:28.903 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:28.904 19:13:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:29.844 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 2 00:17:29.844 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:29.845 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:30.417 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:30.417 { 00:17:30.417 "cntlid": 37, 00:17:30.417 "qid": 0, 00:17:30.417 "state": "enabled", 00:17:30.417 "thread": "nvmf_tgt_poll_group_000", 00:17:30.417 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:30.417 "listen_address": { 00:17:30.417 "trtype": "TCP", 00:17:30.417 "adrfam": "IPv4", 00:17:30.417 "traddr": "10.0.0.2", 00:17:30.417 "trsvcid": "4420" 00:17:30.417 }, 00:17:30.417 "peer_address": { 00:17:30.417 "trtype": "TCP", 00:17:30.417 "adrfam": "IPv4", 00:17:30.417 "traddr": "10.0.0.1", 00:17:30.417 "trsvcid": "54718" 00:17:30.417 }, 00:17:30.417 "auth": { 00:17:30.417 "state": "completed", 00:17:30.417 "digest": "sha256", 00:17:30.417 "dhgroup": "ffdhe6144" 00:17:30.417 } 00:17:30.417 } 00:17:30.417 ]' 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:30.417 19:13:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:30.417 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:30.677 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:30.677 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:30.677 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:30.677 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:30.677 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:30.677 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:30.677 19:13:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:31.619 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 3 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:31.619 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:31.620 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:31.620 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:31.620 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:31.620 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.881 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:31.881 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:31.881 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:31.881 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:32.141 00:17:32.141 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:32.141 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:32.141 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:32.402 { 00:17:32.402 "cntlid": 39, 00:17:32.402 "qid": 0, 00:17:32.402 "state": "enabled", 00:17:32.402 "thread": "nvmf_tgt_poll_group_000", 00:17:32.402 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:32.402 "listen_address": { 00:17:32.402 "trtype": "TCP", 00:17:32.402 "adrfam": "IPv4", 00:17:32.402 "traddr": "10.0.0.2", 00:17:32.402 "trsvcid": "4420" 00:17:32.402 }, 00:17:32.402 "peer_address": { 00:17:32.402 "trtype": "TCP", 00:17:32.402 "adrfam": "IPv4", 00:17:32.402 "traddr": "10.0.0.1", 00:17:32.402 "trsvcid": "54744" 00:17:32.402 }, 00:17:32.402 "auth": { 00:17:32.402 "state": "completed", 00:17:32.402 "digest": "sha256", 00:17:32.402 "dhgroup": "ffdhe6144" 00:17:32.402 } 00:17:32.402 } 00:17:32.402 ]' 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:32.402 19:13:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:32.662 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:32.662 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:33.233 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:33.494 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:33.494 19:13:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 0 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:33.494 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:34.063 00:17:34.063 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:34.063 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:34.063 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:34.324 { 00:17:34.324 "cntlid": 41, 00:17:34.324 "qid": 0, 00:17:34.324 "state": "enabled", 00:17:34.324 "thread": "nvmf_tgt_poll_group_000", 00:17:34.324 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:34.324 "listen_address": { 00:17:34.324 "trtype": "TCP", 00:17:34.324 "adrfam": "IPv4", 00:17:34.324 "traddr": "10.0.0.2", 00:17:34.324 "trsvcid": "4420" 00:17:34.324 }, 00:17:34.324 "peer_address": { 00:17:34.324 "trtype": "TCP", 00:17:34.324 "adrfam": "IPv4", 00:17:34.324 "traddr": "10.0.0.1", 00:17:34.324 "trsvcid": "54776" 00:17:34.324 }, 00:17:34.324 "auth": { 00:17:34.324 "state": "completed", 00:17:34.324 "digest": "sha256", 00:17:34.324 "dhgroup": "ffdhe8192" 00:17:34.324 } 00:17:34.324 } 00:17:34.324 ]' 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:34.324 19:13:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:34.589 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:34.589 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:35.531 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:35.531 19:13:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 1 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:35.531 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:36.102 00:17:36.102 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:36.102 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:36.102 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:36.362 { 00:17:36.362 "cntlid": 43, 00:17:36.362 "qid": 0, 00:17:36.362 "state": "enabled", 00:17:36.362 "thread": "nvmf_tgt_poll_group_000", 00:17:36.362 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:36.362 "listen_address": { 00:17:36.362 "trtype": "TCP", 00:17:36.362 "adrfam": "IPv4", 00:17:36.362 "traddr": "10.0.0.2", 00:17:36.362 "trsvcid": "4420" 00:17:36.362 }, 00:17:36.362 "peer_address": { 00:17:36.362 "trtype": "TCP", 00:17:36.362 "adrfam": "IPv4", 00:17:36.362 "traddr": "10.0.0.1", 00:17:36.362 "trsvcid": "54806" 00:17:36.362 }, 00:17:36.362 "auth": { 00:17:36.362 "state": "completed", 00:17:36.362 "digest": "sha256", 00:17:36.362 "dhgroup": "ffdhe8192" 00:17:36.362 } 00:17:36.362 } 00:17:36.362 ]' 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:36.362 19:13:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:36.623 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:36.623 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:37.562 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:37.562 19:13:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 2 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:37.562 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:38.132 00:17:38.132 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:38.132 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:38.132 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:38.392 { 00:17:38.392 "cntlid": 45, 00:17:38.392 "qid": 0, 00:17:38.392 "state": "enabled", 00:17:38.392 "thread": "nvmf_tgt_poll_group_000", 00:17:38.392 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:38.392 "listen_address": { 00:17:38.392 "trtype": "TCP", 00:17:38.392 "adrfam": "IPv4", 00:17:38.392 "traddr": "10.0.0.2", 00:17:38.392 "trsvcid": "4420" 00:17:38.392 }, 00:17:38.392 "peer_address": { 00:17:38.392 "trtype": "TCP", 00:17:38.392 "adrfam": "IPv4", 00:17:38.392 "traddr": "10.0.0.1", 00:17:38.392 "trsvcid": "54846" 00:17:38.392 }, 00:17:38.392 "auth": { 00:17:38.392 "state": "completed", 00:17:38.392 "digest": "sha256", 00:17:38.392 "dhgroup": "ffdhe8192" 00:17:38.392 } 00:17:38.392 } 00:17:38.392 ]' 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:38.392 19:13:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:38.652 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:38.652 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:39.223 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:39.483 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:39.483 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:39.483 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.483 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:39.483 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.483 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:39.483 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:39.483 19:13:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:17:39.483 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 3 00:17:39.483 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:39.483 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:17:39.483 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:39.483 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:39.483 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:39.483 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:39.484 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.484 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:39.484 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.484 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:39.484 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:39.484 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:40.055 00:17:40.055 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:40.055 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:40.055 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:40.315 { 00:17:40.315 "cntlid": 47, 00:17:40.315 "qid": 0, 00:17:40.315 "state": "enabled", 00:17:40.315 "thread": "nvmf_tgt_poll_group_000", 00:17:40.315 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:40.315 "listen_address": { 00:17:40.315 "trtype": "TCP", 00:17:40.315 "adrfam": "IPv4", 00:17:40.315 "traddr": "10.0.0.2", 00:17:40.315 "trsvcid": "4420" 00:17:40.315 }, 00:17:40.315 "peer_address": { 00:17:40.315 "trtype": "TCP", 00:17:40.315 "adrfam": "IPv4", 00:17:40.315 "traddr": "10.0.0.1", 00:17:40.315 "trsvcid": "37712" 00:17:40.315 }, 00:17:40.315 "auth": { 00:17:40.315 "state": "completed", 00:17:40.315 "digest": "sha256", 00:17:40.315 "dhgroup": "ffdhe8192" 00:17:40.315 } 00:17:40.315 } 00:17:40.315 ]' 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:40.315 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:17:40.316 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:40.316 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:40.316 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:40.316 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:40.316 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:40.316 19:13:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:40.576 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:40.576 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:41.518 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:41.518 19:13:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:41.518 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 0 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:41.519 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:41.780 00:17:41.780 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:41.780 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:41.780 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:42.041 { 00:17:42.041 "cntlid": 49, 00:17:42.041 "qid": 0, 00:17:42.041 "state": "enabled", 00:17:42.041 "thread": "nvmf_tgt_poll_group_000", 00:17:42.041 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:42.041 "listen_address": { 00:17:42.041 "trtype": "TCP", 00:17:42.041 "adrfam": "IPv4", 00:17:42.041 "traddr": "10.0.0.2", 00:17:42.041 "trsvcid": "4420" 00:17:42.041 }, 00:17:42.041 "peer_address": { 00:17:42.041 "trtype": "TCP", 00:17:42.041 "adrfam": "IPv4", 00:17:42.041 "traddr": "10.0.0.1", 00:17:42.041 "trsvcid": "37730" 00:17:42.041 }, 00:17:42.041 "auth": { 00:17:42.041 "state": "completed", 00:17:42.041 "digest": "sha384", 00:17:42.041 "dhgroup": "null" 00:17:42.041 } 00:17:42.041 } 00:17:42.041 ]' 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:42.041 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:42.303 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:42.303 19:13:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:43.245 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 1 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.245 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:43.246 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:43.246 19:13:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:43.507 00:17:43.507 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:43.507 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:43.507 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:43.767 { 00:17:43.767 "cntlid": 51, 00:17:43.767 "qid": 0, 00:17:43.767 "state": "enabled", 00:17:43.767 "thread": "nvmf_tgt_poll_group_000", 00:17:43.767 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:43.767 "listen_address": { 00:17:43.767 "trtype": "TCP", 00:17:43.767 "adrfam": "IPv4", 00:17:43.767 "traddr": "10.0.0.2", 00:17:43.767 "trsvcid": "4420" 00:17:43.767 }, 00:17:43.767 "peer_address": { 00:17:43.767 "trtype": "TCP", 00:17:43.767 "adrfam": "IPv4", 00:17:43.767 "traddr": "10.0.0.1", 00:17:43.767 "trsvcid": "37770" 00:17:43.767 }, 00:17:43.767 "auth": { 00:17:43.767 "state": "completed", 00:17:43.767 "digest": "sha384", 00:17:43.767 "dhgroup": "null" 00:17:43.767 } 00:17:43.767 } 00:17:43.767 ]' 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:43.767 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:44.027 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:44.027 19:13:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:44.969 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 2 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:44.969 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:45.229 00:17:45.229 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:45.229 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:45.229 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:45.489 { 00:17:45.489 "cntlid": 53, 00:17:45.489 "qid": 0, 00:17:45.489 "state": "enabled", 00:17:45.489 "thread": "nvmf_tgt_poll_group_000", 00:17:45.489 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:45.489 "listen_address": { 00:17:45.489 "trtype": "TCP", 00:17:45.489 "adrfam": "IPv4", 00:17:45.489 "traddr": "10.0.0.2", 00:17:45.489 "trsvcid": "4420" 00:17:45.489 }, 00:17:45.489 "peer_address": { 00:17:45.489 "trtype": "TCP", 00:17:45.489 "adrfam": "IPv4", 00:17:45.489 "traddr": "10.0.0.1", 00:17:45.489 "trsvcid": "37806" 00:17:45.489 }, 00:17:45.489 "auth": { 00:17:45.489 "state": "completed", 00:17:45.489 "digest": "sha384", 00:17:45.489 "dhgroup": "null" 00:17:45.489 } 00:17:45.489 } 00:17:45.489 ]' 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:45.489 19:13:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:45.489 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:45.489 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:45.489 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:45.489 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:45.489 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:45.748 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:45.748 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:46.689 19:13:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:46.689 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 3 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:46.689 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:46.949 00:17:46.949 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:46.949 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:46.949 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:47.209 { 00:17:47.209 "cntlid": 55, 00:17:47.209 "qid": 0, 00:17:47.209 "state": "enabled", 00:17:47.209 "thread": "nvmf_tgt_poll_group_000", 00:17:47.209 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:47.209 "listen_address": { 00:17:47.209 "trtype": "TCP", 00:17:47.209 "adrfam": "IPv4", 00:17:47.209 "traddr": "10.0.0.2", 00:17:47.209 "trsvcid": "4420" 00:17:47.209 }, 00:17:47.209 "peer_address": { 00:17:47.209 "trtype": "TCP", 00:17:47.209 "adrfam": "IPv4", 00:17:47.209 "traddr": "10.0.0.1", 00:17:47.209 "trsvcid": "37844" 00:17:47.209 }, 00:17:47.209 "auth": { 00:17:47.209 "state": "completed", 00:17:47.209 "digest": "sha384", 00:17:47.209 "dhgroup": "null" 00:17:47.209 } 00:17:47.209 } 00:17:47.209 ]' 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:47.209 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:47.470 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:47.470 19:13:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:48.410 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 0 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:48.410 19:14:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:48.669 00:17:48.669 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:48.669 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:48.669 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:48.928 { 00:17:48.928 "cntlid": 57, 00:17:48.928 "qid": 0, 00:17:48.928 "state": "enabled", 00:17:48.928 "thread": "nvmf_tgt_poll_group_000", 00:17:48.928 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:48.928 "listen_address": { 00:17:48.928 "trtype": "TCP", 00:17:48.928 "adrfam": "IPv4", 00:17:48.928 "traddr": "10.0.0.2", 00:17:48.928 "trsvcid": "4420" 00:17:48.928 }, 00:17:48.928 "peer_address": { 00:17:48.928 "trtype": "TCP", 00:17:48.928 "adrfam": "IPv4", 00:17:48.928 "traddr": "10.0.0.1", 00:17:48.928 "trsvcid": "37872" 00:17:48.928 }, 00:17:48.928 "auth": { 00:17:48.928 "state": "completed", 00:17:48.928 "digest": "sha384", 00:17:48.928 "dhgroup": "ffdhe2048" 00:17:48.928 } 00:17:48.928 } 00:17:48.928 ]' 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:48.928 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:49.187 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:49.187 19:14:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:50.125 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 1 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:50.125 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:50.385 00:17:50.385 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:50.385 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:50.385 19:14:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:50.645 { 00:17:50.645 "cntlid": 59, 00:17:50.645 "qid": 0, 00:17:50.645 "state": "enabled", 00:17:50.645 "thread": "nvmf_tgt_poll_group_000", 00:17:50.645 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:50.645 "listen_address": { 00:17:50.645 "trtype": "TCP", 00:17:50.645 "adrfam": "IPv4", 00:17:50.645 "traddr": "10.0.0.2", 00:17:50.645 "trsvcid": "4420" 00:17:50.645 }, 00:17:50.645 "peer_address": { 00:17:50.645 "trtype": "TCP", 00:17:50.645 "adrfam": "IPv4", 00:17:50.645 "traddr": "10.0.0.1", 00:17:50.645 "trsvcid": "55454" 00:17:50.645 }, 00:17:50.645 "auth": { 00:17:50.645 "state": "completed", 00:17:50.645 "digest": "sha384", 00:17:50.645 "dhgroup": "ffdhe2048" 00:17:50.645 } 00:17:50.645 } 00:17:50.645 ]' 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:50.645 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:50.905 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:50.905 19:14:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:51.846 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 2 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:51.846 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:52.104 00:17:52.104 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:52.104 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:52.104 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:52.363 { 00:17:52.363 "cntlid": 61, 00:17:52.363 "qid": 0, 00:17:52.363 "state": "enabled", 00:17:52.363 "thread": "nvmf_tgt_poll_group_000", 00:17:52.363 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:52.363 "listen_address": { 00:17:52.363 "trtype": "TCP", 00:17:52.363 "adrfam": "IPv4", 00:17:52.363 "traddr": "10.0.0.2", 00:17:52.363 "trsvcid": "4420" 00:17:52.363 }, 00:17:52.363 "peer_address": { 00:17:52.363 "trtype": "TCP", 00:17:52.363 "adrfam": "IPv4", 00:17:52.363 "traddr": "10.0.0.1", 00:17:52.363 "trsvcid": "55478" 00:17:52.363 }, 00:17:52.363 "auth": { 00:17:52.363 "state": "completed", 00:17:52.363 "digest": "sha384", 00:17:52.363 "dhgroup": "ffdhe2048" 00:17:52.363 } 00:17:52.363 } 00:17:52.363 ]' 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:52.363 19:14:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:52.623 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:52.623 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:53.561 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:53.561 19:14:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 3 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:53.561 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:53.821 00:17:53.821 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:53.821 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:53.821 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:54.081 { 00:17:54.081 "cntlid": 63, 00:17:54.081 "qid": 0, 00:17:54.081 "state": "enabled", 00:17:54.081 "thread": "nvmf_tgt_poll_group_000", 00:17:54.081 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:54.081 "listen_address": { 00:17:54.081 "trtype": "TCP", 00:17:54.081 "adrfam": "IPv4", 00:17:54.081 "traddr": "10.0.0.2", 00:17:54.081 "trsvcid": "4420" 00:17:54.081 }, 00:17:54.081 "peer_address": { 00:17:54.081 "trtype": "TCP", 00:17:54.081 "adrfam": "IPv4", 00:17:54.081 "traddr": "10.0.0.1", 00:17:54.081 "trsvcid": "55500" 00:17:54.081 }, 00:17:54.081 "auth": { 00:17:54.081 "state": "completed", 00:17:54.081 "digest": "sha384", 00:17:54.081 "dhgroup": "ffdhe2048" 00:17:54.081 } 00:17:54.081 } 00:17:54.081 ]' 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:54.081 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:54.341 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:54.341 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:54.341 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:54.341 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:54.341 19:14:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:55.280 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 0 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:55.280 19:14:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:55.540 00:17:55.540 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:55.540 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:55.540 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:55.801 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:55.801 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:55.801 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:55.801 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.801 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:55.801 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:55.801 { 00:17:55.801 "cntlid": 65, 00:17:55.801 "qid": 0, 00:17:55.801 "state": "enabled", 00:17:55.801 "thread": "nvmf_tgt_poll_group_000", 00:17:55.802 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:55.802 "listen_address": { 00:17:55.802 "trtype": "TCP", 00:17:55.802 "adrfam": "IPv4", 00:17:55.802 "traddr": "10.0.0.2", 00:17:55.802 "trsvcid": "4420" 00:17:55.802 }, 00:17:55.802 "peer_address": { 00:17:55.802 "trtype": "TCP", 00:17:55.802 "adrfam": "IPv4", 00:17:55.802 "traddr": "10.0.0.1", 00:17:55.802 "trsvcid": "55514" 00:17:55.802 }, 00:17:55.802 "auth": { 00:17:55.802 "state": "completed", 00:17:55.802 "digest": "sha384", 00:17:55.802 "dhgroup": "ffdhe3072" 00:17:55.802 } 00:17:55.802 } 00:17:55.802 ]' 00:17:55.802 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:55.802 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:55.802 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:55.802 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:55.802 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:56.077 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:56.077 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:56.077 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:56.077 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:56.077 19:14:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:17:57.111 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:57.112 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 1 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:57.112 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:57.373 00:17:57.373 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:57.373 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:57.373 19:14:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:57.634 { 00:17:57.634 "cntlid": 67, 00:17:57.634 "qid": 0, 00:17:57.634 "state": "enabled", 00:17:57.634 "thread": "nvmf_tgt_poll_group_000", 00:17:57.634 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:57.634 "listen_address": { 00:17:57.634 "trtype": "TCP", 00:17:57.634 "adrfam": "IPv4", 00:17:57.634 "traddr": "10.0.0.2", 00:17:57.634 "trsvcid": "4420" 00:17:57.634 }, 00:17:57.634 "peer_address": { 00:17:57.634 "trtype": "TCP", 00:17:57.634 "adrfam": "IPv4", 00:17:57.634 "traddr": "10.0.0.1", 00:17:57.634 "trsvcid": "55536" 00:17:57.634 }, 00:17:57.634 "auth": { 00:17:57.634 "state": "completed", 00:17:57.634 "digest": "sha384", 00:17:57.634 "dhgroup": "ffdhe3072" 00:17:57.634 } 00:17:57.634 } 00:17:57.634 ]' 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:57.634 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:57.894 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:57.894 19:14:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:58.835 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 2 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:58.835 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:59.095 00:17:59.095 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:59.095 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:59.095 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:59.356 { 00:17:59.356 "cntlid": 69, 00:17:59.356 "qid": 0, 00:17:59.356 "state": "enabled", 00:17:59.356 "thread": "nvmf_tgt_poll_group_000", 00:17:59.356 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:17:59.356 "listen_address": { 00:17:59.356 "trtype": "TCP", 00:17:59.356 "adrfam": "IPv4", 00:17:59.356 "traddr": "10.0.0.2", 00:17:59.356 "trsvcid": "4420" 00:17:59.356 }, 00:17:59.356 "peer_address": { 00:17:59.356 "trtype": "TCP", 00:17:59.356 "adrfam": "IPv4", 00:17:59.356 "traddr": "10.0.0.1", 00:17:59.356 "trsvcid": "55554" 00:17:59.356 }, 00:17:59.356 "auth": { 00:17:59.356 "state": "completed", 00:17:59.356 "digest": "sha384", 00:17:59.356 "dhgroup": "ffdhe3072" 00:17:59.356 } 00:17:59.356 } 00:17:59.356 ]' 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:59.356 19:14:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:59.618 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:17:59.618 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:00.558 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:18:00.558 19:14:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 3 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:00.558 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:00.559 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:00.559 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:00.819 00:18:00.819 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:00.819 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:00.819 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:01.080 { 00:18:01.080 "cntlid": 71, 00:18:01.080 "qid": 0, 00:18:01.080 "state": "enabled", 00:18:01.080 "thread": "nvmf_tgt_poll_group_000", 00:18:01.080 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:01.080 "listen_address": { 00:18:01.080 "trtype": "TCP", 00:18:01.080 "adrfam": "IPv4", 00:18:01.080 "traddr": "10.0.0.2", 00:18:01.080 "trsvcid": "4420" 00:18:01.080 }, 00:18:01.080 "peer_address": { 00:18:01.080 "trtype": "TCP", 00:18:01.080 "adrfam": "IPv4", 00:18:01.080 "traddr": "10.0.0.1", 00:18:01.080 "trsvcid": "42786" 00:18:01.080 }, 00:18:01.080 "auth": { 00:18:01.080 "state": "completed", 00:18:01.080 "digest": "sha384", 00:18:01.080 "dhgroup": "ffdhe3072" 00:18:01.080 } 00:18:01.080 } 00:18:01.080 ]' 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:01.080 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:01.340 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:01.340 19:14:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:02.281 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 0 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:02.281 19:14:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:02.542 00:18:02.542 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:02.542 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:02.542 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:02.802 { 00:18:02.802 "cntlid": 73, 00:18:02.802 "qid": 0, 00:18:02.802 "state": "enabled", 00:18:02.802 "thread": "nvmf_tgt_poll_group_000", 00:18:02.802 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:02.802 "listen_address": { 00:18:02.802 "trtype": "TCP", 00:18:02.802 "adrfam": "IPv4", 00:18:02.802 "traddr": "10.0.0.2", 00:18:02.802 "trsvcid": "4420" 00:18:02.802 }, 00:18:02.802 "peer_address": { 00:18:02.802 "trtype": "TCP", 00:18:02.802 "adrfam": "IPv4", 00:18:02.802 "traddr": "10.0.0.1", 00:18:02.802 "trsvcid": "42818" 00:18:02.802 }, 00:18:02.802 "auth": { 00:18:02.802 "state": "completed", 00:18:02.802 "digest": "sha384", 00:18:02.802 "dhgroup": "ffdhe4096" 00:18:02.802 } 00:18:02.802 } 00:18:02.802 ]' 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:02.802 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:02.803 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:03.063 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:03.063 19:14:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:04.027 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:04.027 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:04.027 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:04.027 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:04.027 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:04.027 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:04.027 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:04.027 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 1 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:04.028 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:04.288 00:18:04.288 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:04.288 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:04.288 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:04.549 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:04.549 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:04.549 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:04.549 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:04.549 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:04.549 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:04.549 { 00:18:04.549 "cntlid": 75, 00:18:04.549 "qid": 0, 00:18:04.549 "state": "enabled", 00:18:04.549 "thread": "nvmf_tgt_poll_group_000", 00:18:04.549 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:04.549 "listen_address": { 00:18:04.549 "trtype": "TCP", 00:18:04.549 "adrfam": "IPv4", 00:18:04.549 "traddr": "10.0.0.2", 00:18:04.549 "trsvcid": "4420" 00:18:04.549 }, 00:18:04.549 "peer_address": { 00:18:04.549 "trtype": "TCP", 00:18:04.549 "adrfam": "IPv4", 00:18:04.549 "traddr": "10.0.0.1", 00:18:04.549 "trsvcid": "42842" 00:18:04.549 }, 00:18:04.549 "auth": { 00:18:04.549 "state": "completed", 00:18:04.549 "digest": "sha384", 00:18:04.549 "dhgroup": "ffdhe4096" 00:18:04.549 } 00:18:04.549 } 00:18:04.549 ]' 00:18:04.549 19:14:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:04.549 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:04.549 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:04.549 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:04.549 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:04.549 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:04.549 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:04.549 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:04.809 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:04.809 19:14:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:05.749 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 2 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:05.749 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:06.008 00:18:06.008 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:06.008 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:06.008 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:06.268 { 00:18:06.268 "cntlid": 77, 00:18:06.268 "qid": 0, 00:18:06.268 "state": "enabled", 00:18:06.268 "thread": "nvmf_tgt_poll_group_000", 00:18:06.268 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:06.268 "listen_address": { 00:18:06.268 "trtype": "TCP", 00:18:06.268 "adrfam": "IPv4", 00:18:06.268 "traddr": "10.0.0.2", 00:18:06.268 "trsvcid": "4420" 00:18:06.268 }, 00:18:06.268 "peer_address": { 00:18:06.268 "trtype": "TCP", 00:18:06.268 "adrfam": "IPv4", 00:18:06.268 "traddr": "10.0.0.1", 00:18:06.268 "trsvcid": "42876" 00:18:06.268 }, 00:18:06.268 "auth": { 00:18:06.268 "state": "completed", 00:18:06.268 "digest": "sha384", 00:18:06.268 "dhgroup": "ffdhe4096" 00:18:06.268 } 00:18:06.268 } 00:18:06.268 ]' 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:06.268 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:06.528 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:06.528 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:06.528 19:14:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:06.528 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:06.528 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:07.469 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:07.469 19:14:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 3 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:07.469 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:07.730 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:07.989 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:07.989 { 00:18:07.990 "cntlid": 79, 00:18:07.990 "qid": 0, 00:18:07.990 "state": "enabled", 00:18:07.990 "thread": "nvmf_tgt_poll_group_000", 00:18:07.990 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:07.990 "listen_address": { 00:18:07.990 "trtype": "TCP", 00:18:07.990 "adrfam": "IPv4", 00:18:07.990 "traddr": "10.0.0.2", 00:18:07.990 "trsvcid": "4420" 00:18:07.990 }, 00:18:07.990 "peer_address": { 00:18:07.990 "trtype": "TCP", 00:18:07.990 "adrfam": "IPv4", 00:18:07.990 "traddr": "10.0.0.1", 00:18:07.990 "trsvcid": "42902" 00:18:07.990 }, 00:18:07.990 "auth": { 00:18:07.990 "state": "completed", 00:18:07.990 "digest": "sha384", 00:18:07.990 "dhgroup": "ffdhe4096" 00:18:07.990 } 00:18:07.990 } 00:18:07.990 ]' 00:18:07.990 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:07.990 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:07.990 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:08.249 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:08.249 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:08.249 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:08.249 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:08.249 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:08.249 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:08.249 19:14:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:09.186 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 0 00:18:09.186 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:09.447 19:14:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:09.707 00:18:09.707 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:09.707 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:09.707 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:09.967 { 00:18:09.967 "cntlid": 81, 00:18:09.967 "qid": 0, 00:18:09.967 "state": "enabled", 00:18:09.967 "thread": "nvmf_tgt_poll_group_000", 00:18:09.967 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:09.967 "listen_address": { 00:18:09.967 "trtype": "TCP", 00:18:09.967 "adrfam": "IPv4", 00:18:09.967 "traddr": "10.0.0.2", 00:18:09.967 "trsvcid": "4420" 00:18:09.967 }, 00:18:09.967 "peer_address": { 00:18:09.967 "trtype": "TCP", 00:18:09.967 "adrfam": "IPv4", 00:18:09.967 "traddr": "10.0.0.1", 00:18:09.967 "trsvcid": "50886" 00:18:09.967 }, 00:18:09.967 "auth": { 00:18:09.967 "state": "completed", 00:18:09.967 "digest": "sha384", 00:18:09.967 "dhgroup": "ffdhe6144" 00:18:09.967 } 00:18:09.967 } 00:18:09.967 ]' 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:09.967 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:10.227 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:10.227 19:14:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:11.166 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 1 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:11.166 19:14:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:11.426 00:18:11.686 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:11.687 { 00:18:11.687 "cntlid": 83, 00:18:11.687 "qid": 0, 00:18:11.687 "state": "enabled", 00:18:11.687 "thread": "nvmf_tgt_poll_group_000", 00:18:11.687 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:11.687 "listen_address": { 00:18:11.687 "trtype": "TCP", 00:18:11.687 "adrfam": "IPv4", 00:18:11.687 "traddr": "10.0.0.2", 00:18:11.687 "trsvcid": "4420" 00:18:11.687 }, 00:18:11.687 "peer_address": { 00:18:11.687 "trtype": "TCP", 00:18:11.687 "adrfam": "IPv4", 00:18:11.687 "traddr": "10.0.0.1", 00:18:11.687 "trsvcid": "50900" 00:18:11.687 }, 00:18:11.687 "auth": { 00:18:11.687 "state": "completed", 00:18:11.687 "digest": "sha384", 00:18:11.687 "dhgroup": "ffdhe6144" 00:18:11.687 } 00:18:11.687 } 00:18:11.687 ]' 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:11.687 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:11.948 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:11.948 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:11.948 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:11.948 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:11.948 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:11.948 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:11.948 19:14:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:12.888 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 2 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:12.888 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:13.148 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:13.148 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:13.148 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:13.148 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:13.408 00:18:13.408 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:13.408 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:13.408 19:14:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:13.668 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:13.669 { 00:18:13.669 "cntlid": 85, 00:18:13.669 "qid": 0, 00:18:13.669 "state": "enabled", 00:18:13.669 "thread": "nvmf_tgt_poll_group_000", 00:18:13.669 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:13.669 "listen_address": { 00:18:13.669 "trtype": "TCP", 00:18:13.669 "adrfam": "IPv4", 00:18:13.669 "traddr": "10.0.0.2", 00:18:13.669 "trsvcid": "4420" 00:18:13.669 }, 00:18:13.669 "peer_address": { 00:18:13.669 "trtype": "TCP", 00:18:13.669 "adrfam": "IPv4", 00:18:13.669 "traddr": "10.0.0.1", 00:18:13.669 "trsvcid": "50934" 00:18:13.669 }, 00:18:13.669 "auth": { 00:18:13.669 "state": "completed", 00:18:13.669 "digest": "sha384", 00:18:13.669 "dhgroup": "ffdhe6144" 00:18:13.669 } 00:18:13.669 } 00:18:13.669 ]' 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:13.669 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:13.929 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:13.929 19:14:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:14.870 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 3 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:14.870 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:15.130 00:18:15.130 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:15.130 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:15.130 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:15.390 { 00:18:15.390 "cntlid": 87, 00:18:15.390 "qid": 0, 00:18:15.390 "state": "enabled", 00:18:15.390 "thread": "nvmf_tgt_poll_group_000", 00:18:15.390 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:15.390 "listen_address": { 00:18:15.390 "trtype": "TCP", 00:18:15.390 "adrfam": "IPv4", 00:18:15.390 "traddr": "10.0.0.2", 00:18:15.390 "trsvcid": "4420" 00:18:15.390 }, 00:18:15.390 "peer_address": { 00:18:15.390 "trtype": "TCP", 00:18:15.390 "adrfam": "IPv4", 00:18:15.390 "traddr": "10.0.0.1", 00:18:15.390 "trsvcid": "50962" 00:18:15.390 }, 00:18:15.390 "auth": { 00:18:15.390 "state": "completed", 00:18:15.390 "digest": "sha384", 00:18:15.390 "dhgroup": "ffdhe6144" 00:18:15.390 } 00:18:15.390 } 00:18:15.390 ]' 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:15.390 19:14:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:15.652 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:15.652 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:15.652 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:15.652 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:15.652 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:16.590 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:16.590 19:14:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:16.590 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 0 00:18:16.590 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:16.590 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:16.590 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:16.591 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:17.158 00:18:17.158 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:17.158 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:17.158 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:17.418 { 00:18:17.418 "cntlid": 89, 00:18:17.418 "qid": 0, 00:18:17.418 "state": "enabled", 00:18:17.418 "thread": "nvmf_tgt_poll_group_000", 00:18:17.418 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:17.418 "listen_address": { 00:18:17.418 "trtype": "TCP", 00:18:17.418 "adrfam": "IPv4", 00:18:17.418 "traddr": "10.0.0.2", 00:18:17.418 "trsvcid": "4420" 00:18:17.418 }, 00:18:17.418 "peer_address": { 00:18:17.418 "trtype": "TCP", 00:18:17.418 "adrfam": "IPv4", 00:18:17.418 "traddr": "10.0.0.1", 00:18:17.418 "trsvcid": "51006" 00:18:17.418 }, 00:18:17.418 "auth": { 00:18:17.418 "state": "completed", 00:18:17.418 "digest": "sha384", 00:18:17.418 "dhgroup": "ffdhe8192" 00:18:17.418 } 00:18:17.418 } 00:18:17.418 ]' 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:17.418 19:14:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:17.678 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:17.678 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:18.617 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:18.617 19:14:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 1 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:18.617 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:19.186 00:18:19.186 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:19.186 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:19.186 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:19.445 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:19.445 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:19.445 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:19.445 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:19.445 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:19.445 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:19.445 { 00:18:19.445 "cntlid": 91, 00:18:19.445 "qid": 0, 00:18:19.445 "state": "enabled", 00:18:19.445 "thread": "nvmf_tgt_poll_group_000", 00:18:19.445 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:19.445 "listen_address": { 00:18:19.445 "trtype": "TCP", 00:18:19.445 "adrfam": "IPv4", 00:18:19.445 "traddr": "10.0.0.2", 00:18:19.445 "trsvcid": "4420" 00:18:19.445 }, 00:18:19.445 "peer_address": { 00:18:19.445 "trtype": "TCP", 00:18:19.445 "adrfam": "IPv4", 00:18:19.445 "traddr": "10.0.0.1", 00:18:19.445 "trsvcid": "51028" 00:18:19.445 }, 00:18:19.446 "auth": { 00:18:19.446 "state": "completed", 00:18:19.446 "digest": "sha384", 00:18:19.446 "dhgroup": "ffdhe8192" 00:18:19.446 } 00:18:19.446 } 00:18:19.446 ]' 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:19.446 19:14:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:19.705 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:19.705 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:20.645 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:20.645 19:14:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 2 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:20.645 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:21.213 00:18:21.213 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:21.213 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:21.213 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:21.473 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:21.473 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:21.473 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:21.473 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:21.473 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:21.473 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:21.473 { 00:18:21.473 "cntlid": 93, 00:18:21.473 "qid": 0, 00:18:21.473 "state": "enabled", 00:18:21.473 "thread": "nvmf_tgt_poll_group_000", 00:18:21.473 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:21.473 "listen_address": { 00:18:21.473 "trtype": "TCP", 00:18:21.473 "adrfam": "IPv4", 00:18:21.474 "traddr": "10.0.0.2", 00:18:21.474 "trsvcid": "4420" 00:18:21.474 }, 00:18:21.474 "peer_address": { 00:18:21.474 "trtype": "TCP", 00:18:21.474 "adrfam": "IPv4", 00:18:21.474 "traddr": "10.0.0.1", 00:18:21.474 "trsvcid": "39966" 00:18:21.474 }, 00:18:21.474 "auth": { 00:18:21.474 "state": "completed", 00:18:21.474 "digest": "sha384", 00:18:21.474 "dhgroup": "ffdhe8192" 00:18:21.474 } 00:18:21.474 } 00:18:21.474 ]' 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:21.474 19:14:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:21.733 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:21.733 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:22.362 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:22.362 19:14:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:18:22.620 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 3 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:22.621 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:23.186 00:18:23.186 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:23.186 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:23.186 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:23.444 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:23.444 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:23.444 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:23.444 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:23.444 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:23.444 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:23.444 { 00:18:23.444 "cntlid": 95, 00:18:23.444 "qid": 0, 00:18:23.444 "state": "enabled", 00:18:23.444 "thread": "nvmf_tgt_poll_group_000", 00:18:23.444 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:23.444 "listen_address": { 00:18:23.444 "trtype": "TCP", 00:18:23.444 "adrfam": "IPv4", 00:18:23.444 "traddr": "10.0.0.2", 00:18:23.444 "trsvcid": "4420" 00:18:23.444 }, 00:18:23.444 "peer_address": { 00:18:23.444 "trtype": "TCP", 00:18:23.444 "adrfam": "IPv4", 00:18:23.445 "traddr": "10.0.0.1", 00:18:23.445 "trsvcid": "39990" 00:18:23.445 }, 00:18:23.445 "auth": { 00:18:23.445 "state": "completed", 00:18:23.445 "digest": "sha384", 00:18:23.445 "dhgroup": "ffdhe8192" 00:18:23.445 } 00:18:23.445 } 00:18:23.445 ]' 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:23.445 19:14:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:23.704 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:23.704 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:24.274 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:24.274 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:24.274 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:24.274 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:24.274 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:24.533 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:24.533 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:18:24.533 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:24.533 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:24.533 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:24.533 19:14:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 0 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:24.533 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:24.534 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:24.534 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:24.793 00:18:24.793 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:24.793 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:24.793 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:25.052 { 00:18:25.052 "cntlid": 97, 00:18:25.052 "qid": 0, 00:18:25.052 "state": "enabled", 00:18:25.052 "thread": "nvmf_tgt_poll_group_000", 00:18:25.052 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:25.052 "listen_address": { 00:18:25.052 "trtype": "TCP", 00:18:25.052 "adrfam": "IPv4", 00:18:25.052 "traddr": "10.0.0.2", 00:18:25.052 "trsvcid": "4420" 00:18:25.052 }, 00:18:25.052 "peer_address": { 00:18:25.052 "trtype": "TCP", 00:18:25.052 "adrfam": "IPv4", 00:18:25.052 "traddr": "10.0.0.1", 00:18:25.052 "trsvcid": "40030" 00:18:25.052 }, 00:18:25.052 "auth": { 00:18:25.052 "state": "completed", 00:18:25.052 "digest": "sha512", 00:18:25.052 "dhgroup": "null" 00:18:25.052 } 00:18:25.052 } 00:18:25.052 ]' 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:25.052 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:25.312 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:25.312 19:14:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:25.883 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:26.144 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 1 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:26.144 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:26.403 00:18:26.403 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:26.403 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:26.403 19:14:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:26.662 { 00:18:26.662 "cntlid": 99, 00:18:26.662 "qid": 0, 00:18:26.662 "state": "enabled", 00:18:26.662 "thread": "nvmf_tgt_poll_group_000", 00:18:26.662 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:26.662 "listen_address": { 00:18:26.662 "trtype": "TCP", 00:18:26.662 "adrfam": "IPv4", 00:18:26.662 "traddr": "10.0.0.2", 00:18:26.662 "trsvcid": "4420" 00:18:26.662 }, 00:18:26.662 "peer_address": { 00:18:26.662 "trtype": "TCP", 00:18:26.662 "adrfam": "IPv4", 00:18:26.662 "traddr": "10.0.0.1", 00:18:26.662 "trsvcid": "40062" 00:18:26.662 }, 00:18:26.662 "auth": { 00:18:26.662 "state": "completed", 00:18:26.662 "digest": "sha512", 00:18:26.662 "dhgroup": "null" 00:18:26.662 } 00:18:26.662 } 00:18:26.662 ]' 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:18:26.662 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:26.921 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:26.921 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:26.921 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:26.921 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:26.921 19:14:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:27.858 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 2 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:27.858 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:27.859 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:27.859 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:27.859 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:28.117 00:18:28.117 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:28.117 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:28.117 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:28.378 { 00:18:28.378 "cntlid": 101, 00:18:28.378 "qid": 0, 00:18:28.378 "state": "enabled", 00:18:28.378 "thread": "nvmf_tgt_poll_group_000", 00:18:28.378 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:28.378 "listen_address": { 00:18:28.378 "trtype": "TCP", 00:18:28.378 "adrfam": "IPv4", 00:18:28.378 "traddr": "10.0.0.2", 00:18:28.378 "trsvcid": "4420" 00:18:28.378 }, 00:18:28.378 "peer_address": { 00:18:28.378 "trtype": "TCP", 00:18:28.378 "adrfam": "IPv4", 00:18:28.378 "traddr": "10.0.0.1", 00:18:28.378 "trsvcid": "40088" 00:18:28.378 }, 00:18:28.378 "auth": { 00:18:28.378 "state": "completed", 00:18:28.378 "digest": "sha512", 00:18:28.378 "dhgroup": "null" 00:18:28.378 } 00:18:28.378 } 00:18:28.378 ]' 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:28.378 19:14:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:28.639 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:28.639 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:29.580 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:29.580 19:14:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 3 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:29.580 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:29.842 00:18:29.842 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:29.842 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:29.842 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:30.103 { 00:18:30.103 "cntlid": 103, 00:18:30.103 "qid": 0, 00:18:30.103 "state": "enabled", 00:18:30.103 "thread": "nvmf_tgt_poll_group_000", 00:18:30.103 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:30.103 "listen_address": { 00:18:30.103 "trtype": "TCP", 00:18:30.103 "adrfam": "IPv4", 00:18:30.103 "traddr": "10.0.0.2", 00:18:30.103 "trsvcid": "4420" 00:18:30.103 }, 00:18:30.103 "peer_address": { 00:18:30.103 "trtype": "TCP", 00:18:30.103 "adrfam": "IPv4", 00:18:30.103 "traddr": "10.0.0.1", 00:18:30.103 "trsvcid": "36104" 00:18:30.103 }, 00:18:30.103 "auth": { 00:18:30.103 "state": "completed", 00:18:30.103 "digest": "sha512", 00:18:30.103 "dhgroup": "null" 00:18:30.103 } 00:18:30.103 } 00:18:30.103 ]' 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:30.103 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:30.363 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:30.363 19:14:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:31.301 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 0 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:31.301 19:14:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:31.560 00:18:31.560 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:31.560 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:31.560 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:31.820 { 00:18:31.820 "cntlid": 105, 00:18:31.820 "qid": 0, 00:18:31.820 "state": "enabled", 00:18:31.820 "thread": "nvmf_tgt_poll_group_000", 00:18:31.820 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:31.820 "listen_address": { 00:18:31.820 "trtype": "TCP", 00:18:31.820 "adrfam": "IPv4", 00:18:31.820 "traddr": "10.0.0.2", 00:18:31.820 "trsvcid": "4420" 00:18:31.820 }, 00:18:31.820 "peer_address": { 00:18:31.820 "trtype": "TCP", 00:18:31.820 "adrfam": "IPv4", 00:18:31.820 "traddr": "10.0.0.1", 00:18:31.820 "trsvcid": "36126" 00:18:31.820 }, 00:18:31.820 "auth": { 00:18:31.820 "state": "completed", 00:18:31.820 "digest": "sha512", 00:18:31.820 "dhgroup": "ffdhe2048" 00:18:31.820 } 00:18:31.820 } 00:18:31.820 ]' 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:31.820 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:32.079 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:32.079 19:14:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:33.018 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 1 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:33.018 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:33.279 00:18:33.279 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:33.279 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:33.279 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:33.539 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:33.539 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:33.539 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:33.539 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:33.539 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:33.539 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:33.539 { 00:18:33.539 "cntlid": 107, 00:18:33.539 "qid": 0, 00:18:33.539 "state": "enabled", 00:18:33.539 "thread": "nvmf_tgt_poll_group_000", 00:18:33.539 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:33.539 "listen_address": { 00:18:33.539 "trtype": "TCP", 00:18:33.539 "adrfam": "IPv4", 00:18:33.539 "traddr": "10.0.0.2", 00:18:33.539 "trsvcid": "4420" 00:18:33.539 }, 00:18:33.539 "peer_address": { 00:18:33.539 "trtype": "TCP", 00:18:33.539 "adrfam": "IPv4", 00:18:33.539 "traddr": "10.0.0.1", 00:18:33.539 "trsvcid": "36160" 00:18:33.539 }, 00:18:33.539 "auth": { 00:18:33.539 "state": "completed", 00:18:33.539 "digest": "sha512", 00:18:33.539 "dhgroup": "ffdhe2048" 00:18:33.539 } 00:18:33.539 } 00:18:33.539 ]' 00:18:33.539 19:14:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:33.539 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:33.539 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:33.539 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:18:33.539 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:33.539 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:33.539 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:33.539 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:33.799 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:33.799 19:14:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:34.793 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:34.793 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:34.793 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:34.793 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:34.793 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 2 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:34.794 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:35.119 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:35.119 { 00:18:35.119 "cntlid": 109, 00:18:35.119 "qid": 0, 00:18:35.119 "state": "enabled", 00:18:35.119 "thread": "nvmf_tgt_poll_group_000", 00:18:35.119 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:35.119 "listen_address": { 00:18:35.119 "trtype": "TCP", 00:18:35.119 "adrfam": "IPv4", 00:18:35.119 "traddr": "10.0.0.2", 00:18:35.119 "trsvcid": "4420" 00:18:35.119 }, 00:18:35.119 "peer_address": { 00:18:35.119 "trtype": "TCP", 00:18:35.119 "adrfam": "IPv4", 00:18:35.119 "traddr": "10.0.0.1", 00:18:35.119 "trsvcid": "36196" 00:18:35.119 }, 00:18:35.119 "auth": { 00:18:35.119 "state": "completed", 00:18:35.119 "digest": "sha512", 00:18:35.119 "dhgroup": "ffdhe2048" 00:18:35.119 } 00:18:35.119 } 00:18:35.119 ]' 00:18:35.119 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:35.407 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:35.407 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:35.407 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:18:35.407 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:35.407 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:35.407 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:35.407 19:14:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:35.407 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:35.407 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:36.349 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:36.349 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 3 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:36.610 19:14:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:36.610 00:18:36.610 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:36.610 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:36.610 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:36.870 { 00:18:36.870 "cntlid": 111, 00:18:36.870 "qid": 0, 00:18:36.870 "state": "enabled", 00:18:36.870 "thread": "nvmf_tgt_poll_group_000", 00:18:36.870 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:36.870 "listen_address": { 00:18:36.870 "trtype": "TCP", 00:18:36.870 "adrfam": "IPv4", 00:18:36.870 "traddr": "10.0.0.2", 00:18:36.870 "trsvcid": "4420" 00:18:36.870 }, 00:18:36.870 "peer_address": { 00:18:36.870 "trtype": "TCP", 00:18:36.870 "adrfam": "IPv4", 00:18:36.870 "traddr": "10.0.0.1", 00:18:36.870 "trsvcid": "36234" 00:18:36.870 }, 00:18:36.870 "auth": { 00:18:36.870 "state": "completed", 00:18:36.870 "digest": "sha512", 00:18:36.870 "dhgroup": "ffdhe2048" 00:18:36.870 } 00:18:36.870 } 00:18:36.870 ]' 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:36.870 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:37.131 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:18:37.131 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:37.131 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:37.131 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:37.131 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:37.131 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:37.131 19:14:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:38.072 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 0 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:38.072 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:38.333 00:18:38.333 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:38.333 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:38.333 19:14:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:38.593 { 00:18:38.593 "cntlid": 113, 00:18:38.593 "qid": 0, 00:18:38.593 "state": "enabled", 00:18:38.593 "thread": "nvmf_tgt_poll_group_000", 00:18:38.593 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:38.593 "listen_address": { 00:18:38.593 "trtype": "TCP", 00:18:38.593 "adrfam": "IPv4", 00:18:38.593 "traddr": "10.0.0.2", 00:18:38.593 "trsvcid": "4420" 00:18:38.593 }, 00:18:38.593 "peer_address": { 00:18:38.593 "trtype": "TCP", 00:18:38.593 "adrfam": "IPv4", 00:18:38.593 "traddr": "10.0.0.1", 00:18:38.593 "trsvcid": "36274" 00:18:38.593 }, 00:18:38.593 "auth": { 00:18:38.593 "state": "completed", 00:18:38.593 "digest": "sha512", 00:18:38.593 "dhgroup": "ffdhe3072" 00:18:38.593 } 00:18:38.593 } 00:18:38.593 ]' 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:38.593 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:38.853 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:18:38.853 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:38.853 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:38.853 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:38.853 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:38.853 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:38.853 19:14:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:39.793 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 1 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:39.793 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:40.053 00:18:40.053 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:40.053 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:40.053 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:40.313 { 00:18:40.313 "cntlid": 115, 00:18:40.313 "qid": 0, 00:18:40.313 "state": "enabled", 00:18:40.313 "thread": "nvmf_tgt_poll_group_000", 00:18:40.313 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:40.313 "listen_address": { 00:18:40.313 "trtype": "TCP", 00:18:40.313 "adrfam": "IPv4", 00:18:40.313 "traddr": "10.0.0.2", 00:18:40.313 "trsvcid": "4420" 00:18:40.313 }, 00:18:40.313 "peer_address": { 00:18:40.313 "trtype": "TCP", 00:18:40.313 "adrfam": "IPv4", 00:18:40.313 "traddr": "10.0.0.1", 00:18:40.313 "trsvcid": "51588" 00:18:40.313 }, 00:18:40.313 "auth": { 00:18:40.313 "state": "completed", 00:18:40.313 "digest": "sha512", 00:18:40.313 "dhgroup": "ffdhe3072" 00:18:40.313 } 00:18:40.313 } 00:18:40.313 ]' 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:40.313 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:40.574 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:18:40.574 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:40.574 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:40.574 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:40.574 19:14:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:40.574 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:40.574 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:41.513 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:41.513 19:14:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 2 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:41.774 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:41.774 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:42.035 { 00:18:42.035 "cntlid": 117, 00:18:42.035 "qid": 0, 00:18:42.035 "state": "enabled", 00:18:42.035 "thread": "nvmf_tgt_poll_group_000", 00:18:42.035 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:42.035 "listen_address": { 00:18:42.035 "trtype": "TCP", 00:18:42.035 "adrfam": "IPv4", 00:18:42.035 "traddr": "10.0.0.2", 00:18:42.035 "trsvcid": "4420" 00:18:42.035 }, 00:18:42.035 "peer_address": { 00:18:42.035 "trtype": "TCP", 00:18:42.035 "adrfam": "IPv4", 00:18:42.035 "traddr": "10.0.0.1", 00:18:42.035 "trsvcid": "51616" 00:18:42.035 }, 00:18:42.035 "auth": { 00:18:42.035 "state": "completed", 00:18:42.035 "digest": "sha512", 00:18:42.035 "dhgroup": "ffdhe3072" 00:18:42.035 } 00:18:42.035 } 00:18:42.035 ]' 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:42.035 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:42.296 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:18:42.296 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:42.296 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:42.296 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:42.296 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:42.556 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:42.556 19:14:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:43.127 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:43.388 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 3 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:43.388 19:14:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:43.648 00:18:43.648 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:43.648 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:43.648 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:43.909 { 00:18:43.909 "cntlid": 119, 00:18:43.909 "qid": 0, 00:18:43.909 "state": "enabled", 00:18:43.909 "thread": "nvmf_tgt_poll_group_000", 00:18:43.909 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:43.909 "listen_address": { 00:18:43.909 "trtype": "TCP", 00:18:43.909 "adrfam": "IPv4", 00:18:43.909 "traddr": "10.0.0.2", 00:18:43.909 "trsvcid": "4420" 00:18:43.909 }, 00:18:43.909 "peer_address": { 00:18:43.909 "trtype": "TCP", 00:18:43.909 "adrfam": "IPv4", 00:18:43.909 "traddr": "10.0.0.1", 00:18:43.909 "trsvcid": "51648" 00:18:43.909 }, 00:18:43.909 "auth": { 00:18:43.909 "state": "completed", 00:18:43.909 "digest": "sha512", 00:18:43.909 "dhgroup": "ffdhe3072" 00:18:43.909 } 00:18:43.909 } 00:18:43.909 ]' 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:18:43.909 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:44.170 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:44.170 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:44.170 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:44.170 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:44.170 19:14:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:45.113 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 0 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:45.113 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:45.375 00:18:45.635 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:45.635 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:45.636 19:14:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:45.636 { 00:18:45.636 "cntlid": 121, 00:18:45.636 "qid": 0, 00:18:45.636 "state": "enabled", 00:18:45.636 "thread": "nvmf_tgt_poll_group_000", 00:18:45.636 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:45.636 "listen_address": { 00:18:45.636 "trtype": "TCP", 00:18:45.636 "adrfam": "IPv4", 00:18:45.636 "traddr": "10.0.0.2", 00:18:45.636 "trsvcid": "4420" 00:18:45.636 }, 00:18:45.636 "peer_address": { 00:18:45.636 "trtype": "TCP", 00:18:45.636 "adrfam": "IPv4", 00:18:45.636 "traddr": "10.0.0.1", 00:18:45.636 "trsvcid": "51656" 00:18:45.636 }, 00:18:45.636 "auth": { 00:18:45.636 "state": "completed", 00:18:45.636 "digest": "sha512", 00:18:45.636 "dhgroup": "ffdhe4096" 00:18:45.636 } 00:18:45.636 } 00:18:45.636 ]' 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:45.636 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:45.896 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:45.896 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:45.896 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:45.896 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:45.896 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:45.896 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:45.896 19:14:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:46.837 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:46.837 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 1 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:47.097 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:47.357 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:47.357 { 00:18:47.357 "cntlid": 123, 00:18:47.357 "qid": 0, 00:18:47.357 "state": "enabled", 00:18:47.357 "thread": "nvmf_tgt_poll_group_000", 00:18:47.357 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:47.357 "listen_address": { 00:18:47.357 "trtype": "TCP", 00:18:47.357 "adrfam": "IPv4", 00:18:47.357 "traddr": "10.0.0.2", 00:18:47.357 "trsvcid": "4420" 00:18:47.357 }, 00:18:47.357 "peer_address": { 00:18:47.357 "trtype": "TCP", 00:18:47.357 "adrfam": "IPv4", 00:18:47.357 "traddr": "10.0.0.1", 00:18:47.357 "trsvcid": "51674" 00:18:47.357 }, 00:18:47.357 "auth": { 00:18:47.357 "state": "completed", 00:18:47.357 "digest": "sha512", 00:18:47.357 "dhgroup": "ffdhe4096" 00:18:47.357 } 00:18:47.357 } 00:18:47.357 ]' 00:18:47.357 19:14:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:47.618 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:47.618 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:47.618 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:47.618 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:47.618 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:47.618 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:47.618 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:47.878 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:47.878 19:15:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:48.449 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:48.449 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 2 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:48.708 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:48.969 00:18:48.969 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:48.969 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:48.969 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:49.229 { 00:18:49.229 "cntlid": 125, 00:18:49.229 "qid": 0, 00:18:49.229 "state": "enabled", 00:18:49.229 "thread": "nvmf_tgt_poll_group_000", 00:18:49.229 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:49.229 "listen_address": { 00:18:49.229 "trtype": "TCP", 00:18:49.229 "adrfam": "IPv4", 00:18:49.229 "traddr": "10.0.0.2", 00:18:49.229 "trsvcid": "4420" 00:18:49.229 }, 00:18:49.229 "peer_address": { 00:18:49.229 "trtype": "TCP", 00:18:49.229 "adrfam": "IPv4", 00:18:49.229 "traddr": "10.0.0.1", 00:18:49.229 "trsvcid": "51706" 00:18:49.229 }, 00:18:49.229 "auth": { 00:18:49.229 "state": "completed", 00:18:49.229 "digest": "sha512", 00:18:49.229 "dhgroup": "ffdhe4096" 00:18:49.229 } 00:18:49.229 } 00:18:49.229 ]' 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:49.229 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:49.489 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:49.489 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:49.489 19:15:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:49.489 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:49.489 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:50.428 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:18:50.428 19:15:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 3 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:50.428 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:50.689 00:18:50.689 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:50.689 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:50.689 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:50.948 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:50.949 { 00:18:50.949 "cntlid": 127, 00:18:50.949 "qid": 0, 00:18:50.949 "state": "enabled", 00:18:50.949 "thread": "nvmf_tgt_poll_group_000", 00:18:50.949 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:50.949 "listen_address": { 00:18:50.949 "trtype": "TCP", 00:18:50.949 "adrfam": "IPv4", 00:18:50.949 "traddr": "10.0.0.2", 00:18:50.949 "trsvcid": "4420" 00:18:50.949 }, 00:18:50.949 "peer_address": { 00:18:50.949 "trtype": "TCP", 00:18:50.949 "adrfam": "IPv4", 00:18:50.949 "traddr": "10.0.0.1", 00:18:50.949 "trsvcid": "40050" 00:18:50.949 }, 00:18:50.949 "auth": { 00:18:50.949 "state": "completed", 00:18:50.949 "digest": "sha512", 00:18:50.949 "dhgroup": "ffdhe4096" 00:18:50.949 } 00:18:50.949 } 00:18:50.949 ]' 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:50.949 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:51.209 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:18:51.209 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:51.210 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:51.210 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:51.210 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:51.210 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:51.210 19:15:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:52.151 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 0 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:52.151 19:15:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:52.723 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:52.723 { 00:18:52.723 "cntlid": 129, 00:18:52.723 "qid": 0, 00:18:52.723 "state": "enabled", 00:18:52.723 "thread": "nvmf_tgt_poll_group_000", 00:18:52.723 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:52.723 "listen_address": { 00:18:52.723 "trtype": "TCP", 00:18:52.723 "adrfam": "IPv4", 00:18:52.723 "traddr": "10.0.0.2", 00:18:52.723 "trsvcid": "4420" 00:18:52.723 }, 00:18:52.723 "peer_address": { 00:18:52.723 "trtype": "TCP", 00:18:52.723 "adrfam": "IPv4", 00:18:52.723 "traddr": "10.0.0.1", 00:18:52.723 "trsvcid": "40094" 00:18:52.723 }, 00:18:52.723 "auth": { 00:18:52.723 "state": "completed", 00:18:52.723 "digest": "sha512", 00:18:52.723 "dhgroup": "ffdhe6144" 00:18:52.723 } 00:18:52.723 } 00:18:52.723 ]' 00:18:52.723 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:52.983 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:52.983 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:52.983 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:52.983 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:52.983 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:52.983 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:52.983 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:53.244 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:53.244 19:15:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:53.815 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:53.815 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 1 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:54.076 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:18:54.337 00:18:54.597 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:54.597 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:54.597 19:15:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:54.597 { 00:18:54.597 "cntlid": 131, 00:18:54.597 "qid": 0, 00:18:54.597 "state": "enabled", 00:18:54.597 "thread": "nvmf_tgt_poll_group_000", 00:18:54.597 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:54.597 "listen_address": { 00:18:54.597 "trtype": "TCP", 00:18:54.597 "adrfam": "IPv4", 00:18:54.597 "traddr": "10.0.0.2", 00:18:54.597 "trsvcid": "4420" 00:18:54.597 }, 00:18:54.597 "peer_address": { 00:18:54.597 "trtype": "TCP", 00:18:54.597 "adrfam": "IPv4", 00:18:54.597 "traddr": "10.0.0.1", 00:18:54.597 "trsvcid": "40124" 00:18:54.597 }, 00:18:54.597 "auth": { 00:18:54.597 "state": "completed", 00:18:54.597 "digest": "sha512", 00:18:54.597 "dhgroup": "ffdhe6144" 00:18:54.597 } 00:18:54.597 } 00:18:54.597 ]' 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:54.597 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:54.857 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:54.857 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:54.857 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:54.857 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:54.857 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:54.857 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:54.857 19:15:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:55.800 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 2 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:55.800 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:56.060 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:56.060 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:56.060 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:56.060 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:18:56.320 00:18:56.320 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:56.320 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:56.320 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:56.579 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:56.579 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:56.579 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:56.579 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:56.579 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:56.579 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:56.579 { 00:18:56.579 "cntlid": 133, 00:18:56.579 "qid": 0, 00:18:56.579 "state": "enabled", 00:18:56.579 "thread": "nvmf_tgt_poll_group_000", 00:18:56.579 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:56.579 "listen_address": { 00:18:56.579 "trtype": "TCP", 00:18:56.579 "adrfam": "IPv4", 00:18:56.579 "traddr": "10.0.0.2", 00:18:56.579 "trsvcid": "4420" 00:18:56.579 }, 00:18:56.579 "peer_address": { 00:18:56.579 "trtype": "TCP", 00:18:56.579 "adrfam": "IPv4", 00:18:56.579 "traddr": "10.0.0.1", 00:18:56.579 "trsvcid": "40152" 00:18:56.579 }, 00:18:56.579 "auth": { 00:18:56.579 "state": "completed", 00:18:56.579 "digest": "sha512", 00:18:56.579 "dhgroup": "ffdhe6144" 00:18:56.579 } 00:18:56.579 } 00:18:56.579 ]' 00:18:56.579 19:15:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:56.579 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:56.579 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:56.579 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:56.579 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:56.579 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:56.579 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:56.579 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:56.838 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:56.838 19:15:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:18:57.777 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:57.777 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:57.777 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:57.777 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:57.777 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 3 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:57.778 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:58.038 00:18:58.038 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:18:58.038 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:18:58.038 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:18:58.297 { 00:18:58.297 "cntlid": 135, 00:18:58.297 "qid": 0, 00:18:58.297 "state": "enabled", 00:18:58.297 "thread": "nvmf_tgt_poll_group_000", 00:18:58.297 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:18:58.297 "listen_address": { 00:18:58.297 "trtype": "TCP", 00:18:58.297 "adrfam": "IPv4", 00:18:58.297 "traddr": "10.0.0.2", 00:18:58.297 "trsvcid": "4420" 00:18:58.297 }, 00:18:58.297 "peer_address": { 00:18:58.297 "trtype": "TCP", 00:18:58.297 "adrfam": "IPv4", 00:18:58.297 "traddr": "10.0.0.1", 00:18:58.297 "trsvcid": "40174" 00:18:58.297 }, 00:18:58.297 "auth": { 00:18:58.297 "state": "completed", 00:18:58.297 "digest": "sha512", 00:18:58.297 "dhgroup": "ffdhe6144" 00:18:58.297 } 00:18:58.297 } 00:18:58.297 ]' 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:18:58.297 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:18:58.558 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:18:58.558 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:18:58.558 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:18:58.558 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:58.558 19:15:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:58.558 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:58.558 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:59.499 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:18:59.499 19:15:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 0 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:18:59.759 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:19:00.329 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:19:00.329 { 00:19:00.329 "cntlid": 137, 00:19:00.329 "qid": 0, 00:19:00.329 "state": "enabled", 00:19:00.329 "thread": "nvmf_tgt_poll_group_000", 00:19:00.329 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:00.329 "listen_address": { 00:19:00.329 "trtype": "TCP", 00:19:00.329 "adrfam": "IPv4", 00:19:00.329 "traddr": "10.0.0.2", 00:19:00.329 "trsvcid": "4420" 00:19:00.329 }, 00:19:00.329 "peer_address": { 00:19:00.329 "trtype": "TCP", 00:19:00.329 "adrfam": "IPv4", 00:19:00.329 "traddr": "10.0.0.1", 00:19:00.329 "trsvcid": "40856" 00:19:00.329 }, 00:19:00.329 "auth": { 00:19:00.329 "state": "completed", 00:19:00.329 "digest": "sha512", 00:19:00.329 "dhgroup": "ffdhe8192" 00:19:00.329 } 00:19:00.329 } 00:19:00.329 ]' 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:19:00.329 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:19:00.590 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:19:00.590 19:15:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:19:00.590 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:19:00.590 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:00.590 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:00.590 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:19:00.590 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:19:01.532 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:19:01.532 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:19:01.532 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:01.532 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:01.532 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:01.533 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:01.533 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:19:01.533 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:19:01.533 19:15:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 1 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:01.533 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:02.105 00:19:02.105 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:19:02.105 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:19:02.105 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:19:02.366 { 00:19:02.366 "cntlid": 139, 00:19:02.366 "qid": 0, 00:19:02.366 "state": "enabled", 00:19:02.366 "thread": "nvmf_tgt_poll_group_000", 00:19:02.366 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:02.366 "listen_address": { 00:19:02.366 "trtype": "TCP", 00:19:02.366 "adrfam": "IPv4", 00:19:02.366 "traddr": "10.0.0.2", 00:19:02.366 "trsvcid": "4420" 00:19:02.366 }, 00:19:02.366 "peer_address": { 00:19:02.366 "trtype": "TCP", 00:19:02.366 "adrfam": "IPv4", 00:19:02.366 "traddr": "10.0.0.1", 00:19:02.366 "trsvcid": "40882" 00:19:02.366 }, 00:19:02.366 "auth": { 00:19:02.366 "state": "completed", 00:19:02.366 "digest": "sha512", 00:19:02.366 "dhgroup": "ffdhe8192" 00:19:02.366 } 00:19:02.366 } 00:19:02.366 ]' 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:19:02.366 19:15:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:19:02.626 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:19:02.627 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:02.627 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:02.627 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:19:02.627 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: --dhchap-ctrl-secret DHHC-1:02:MjdmZjRmNzllZTFkNTcxMWU4MzJjZTM2NDJkMmU0YTBhYTRiNmMxOTE0YThjNDAyKU6FmQ==: 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:19:03.569 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:19:03.569 19:15:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 2 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:19:03.569 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:19:04.140 00:19:04.140 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:19:04.140 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:19:04.140 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:19:04.401 { 00:19:04.401 "cntlid": 141, 00:19:04.401 "qid": 0, 00:19:04.401 "state": "enabled", 00:19:04.401 "thread": "nvmf_tgt_poll_group_000", 00:19:04.401 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:04.401 "listen_address": { 00:19:04.401 "trtype": "TCP", 00:19:04.401 "adrfam": "IPv4", 00:19:04.401 "traddr": "10.0.0.2", 00:19:04.401 "trsvcid": "4420" 00:19:04.401 }, 00:19:04.401 "peer_address": { 00:19:04.401 "trtype": "TCP", 00:19:04.401 "adrfam": "IPv4", 00:19:04.401 "traddr": "10.0.0.1", 00:19:04.401 "trsvcid": "40904" 00:19:04.401 }, 00:19:04.401 "auth": { 00:19:04.401 "state": "completed", 00:19:04.401 "digest": "sha512", 00:19:04.401 "dhgroup": "ffdhe8192" 00:19:04.401 } 00:19:04.401 } 00:19:04.401 ]' 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:19:04.401 19:15:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:19:04.662 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:19:04.662 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:04.662 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:04.662 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:19:04.662 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:01:M2VjMDk4Yjk3ZDFiYzA2N2IzYWY1MWUyZjY2ODViZGGclS1p: 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:19:05.603 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:19:05.603 19:15:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 3 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:05.603 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:06.175 00:19:06.175 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:19:06.175 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:19:06.175 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:06.436 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:06.436 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:19:06.436 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:06.436 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:06.436 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:06.436 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:19:06.436 { 00:19:06.436 "cntlid": 143, 00:19:06.436 "qid": 0, 00:19:06.436 "state": "enabled", 00:19:06.436 "thread": "nvmf_tgt_poll_group_000", 00:19:06.436 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:06.436 "listen_address": { 00:19:06.436 "trtype": "TCP", 00:19:06.436 "adrfam": "IPv4", 00:19:06.436 "traddr": "10.0.0.2", 00:19:06.436 "trsvcid": "4420" 00:19:06.436 }, 00:19:06.436 "peer_address": { 00:19:06.436 "trtype": "TCP", 00:19:06.436 "adrfam": "IPv4", 00:19:06.436 "traddr": "10.0.0.1", 00:19:06.436 "trsvcid": "40938" 00:19:06.436 }, 00:19:06.436 "auth": { 00:19:06.436 "state": "completed", 00:19:06.436 "digest": "sha512", 00:19:06.436 "dhgroup": "ffdhe8192" 00:19:06.436 } 00:19:06.436 } 00:19:06.436 ]' 00:19:06.436 19:15:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:19:06.436 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:19:06.436 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:19:06.436 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:19:06.436 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:19:06.697 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:19:06.697 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:06.697 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:06.697 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:19:06.697 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:19:07.638 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:19:07.638 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:19:07.638 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:07.638 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:07.638 19:15:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s sha256,sha384,sha512 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@141 -- # connect_authenticate sha512 ffdhe8192 0 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:19:07.638 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:19:08.209 00:19:08.209 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:19:08.209 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:19:08.209 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:19:08.469 { 00:19:08.469 "cntlid": 145, 00:19:08.469 "qid": 0, 00:19:08.469 "state": "enabled", 00:19:08.469 "thread": "nvmf_tgt_poll_group_000", 00:19:08.469 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:08.469 "listen_address": { 00:19:08.469 "trtype": "TCP", 00:19:08.469 "adrfam": "IPv4", 00:19:08.469 "traddr": "10.0.0.2", 00:19:08.469 "trsvcid": "4420" 00:19:08.469 }, 00:19:08.469 "peer_address": { 00:19:08.469 "trtype": "TCP", 00:19:08.469 "adrfam": "IPv4", 00:19:08.469 "traddr": "10.0.0.1", 00:19:08.469 "trsvcid": "40954" 00:19:08.469 }, 00:19:08.469 "auth": { 00:19:08.469 "state": "completed", 00:19:08.469 "digest": "sha512", 00:19:08.469 "dhgroup": "ffdhe8192" 00:19:08.469 } 00:19:08.469 } 00:19:08.469 ]' 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:19:08.469 19:15:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:19:08.469 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:19:08.469 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:19:08.469 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:19:08.469 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:08.469 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:08.731 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:19:08.731 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:00:NjA3ODVkNjlhZmRjMzA2YTU5YzE1NzFmYTQ0Yjc3N2E1MGM1MDhlYWVlYjZkMTMy/I+vbw==: --dhchap-ctrl-secret DHHC-1:03:OTBhMjRjYjcwYzBhYjQwMGNkZDRhYzVhYzJmZGE5YzM0OTEyZjcwNzI5ODY3OWRjNTdhOWQyNzFmOTJhMzEwNPG/nKU=: 00:19:09.674 19:15:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:19:09.674 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@144 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@145 -- # NOT bdev_connect -b nvme0 --dhchap-key key2 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key2 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key2 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:19:09.675 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:19:09.936 request: 00:19:09.936 { 00:19:09.936 "name": "nvme0", 00:19:09.936 "trtype": "tcp", 00:19:09.936 "traddr": "10.0.0.2", 00:19:09.936 "adrfam": "ipv4", 00:19:09.936 "trsvcid": "4420", 00:19:09.936 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:19:09.936 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:09.936 "prchk_reftag": false, 00:19:09.936 "prchk_guard": false, 00:19:09.936 "hdgst": false, 00:19:09.936 "ddgst": false, 00:19:09.936 "dhchap_key": "key2", 00:19:09.936 "allow_unrecognized_csi": false, 00:19:09.936 "method": "bdev_nvme_attach_controller", 00:19:09.936 "req_id": 1 00:19:09.936 } 00:19:09.936 Got JSON-RPC error response 00:19:09.936 response: 00:19:09.936 { 00:19:09.936 "code": -5, 00:19:09.936 "message": "Input/output error" 00:19:09.936 } 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@146 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@149 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:09.936 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@150 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:19:10.197 19:15:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:19:10.457 request: 00:19:10.457 { 00:19:10.457 "name": "nvme0", 00:19:10.457 "trtype": "tcp", 00:19:10.457 "traddr": "10.0.0.2", 00:19:10.457 "adrfam": "ipv4", 00:19:10.457 "trsvcid": "4420", 00:19:10.457 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:19:10.457 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:10.457 "prchk_reftag": false, 00:19:10.457 "prchk_guard": false, 00:19:10.457 "hdgst": false, 00:19:10.457 "ddgst": false, 00:19:10.457 "dhchap_key": "key1", 00:19:10.457 "dhchap_ctrlr_key": "ckey2", 00:19:10.457 "allow_unrecognized_csi": false, 00:19:10.457 "method": "bdev_nvme_attach_controller", 00:19:10.457 "req_id": 1 00:19:10.457 } 00:19:10.457 Got JSON-RPC error response 00:19:10.457 response: 00:19:10.457 { 00:19:10.457 "code": -5, 00:19:10.457 "message": "Input/output error" 00:19:10.457 } 00:19:10.457 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:10.457 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:10.457 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:10.457 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:10.457 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@151 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:10.457 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:10.457 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@154 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@155 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:10.718 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:19:10.979 request: 00:19:10.979 { 00:19:10.979 "name": "nvme0", 00:19:10.979 "trtype": "tcp", 00:19:10.979 "traddr": "10.0.0.2", 00:19:10.979 "adrfam": "ipv4", 00:19:10.979 "trsvcid": "4420", 00:19:10.979 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:19:10.979 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:10.979 "prchk_reftag": false, 00:19:10.979 "prchk_guard": false, 00:19:10.979 "hdgst": false, 00:19:10.979 "ddgst": false, 00:19:10.979 "dhchap_key": "key1", 00:19:10.979 "dhchap_ctrlr_key": "ckey1", 00:19:10.979 "allow_unrecognized_csi": false, 00:19:10.979 "method": "bdev_nvme_attach_controller", 00:19:10.979 "req_id": 1 00:19:10.979 } 00:19:10.979 Got JSON-RPC error response 00:19:10.979 response: 00:19:10.979 { 00:19:10.979 "code": -5, 00:19:10.979 "message": "Input/output error" 00:19:10.979 } 00:19:10.979 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:10.979 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:10.979 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:10.979 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:10.979 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@156 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:10.979 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:10.979 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@159 -- # killprocess 3828270 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # '[' -z 3828270 ']' 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@958 -- # kill -0 3828270 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # uname 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3828270 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3828270' 00:19:11.240 killing process with pid 3828270 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@973 -- # kill 3828270 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@978 -- # wait 3828270 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@160 -- # nvmfappstart --wait-for-rpc -L nvmf_auth 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # nvmfpid=3856531 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # waitforlisten 3856531 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc -L nvmf_auth 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 3856531 ']' 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:11.240 19:15:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.182 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:12.182 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@161 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@163 -- # waitforlisten 3856531 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 3856531 ']' 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:12.183 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:12.183 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@164 -- # rpc_cmd 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.443 null0 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.uaB 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha512.tl0 ]] 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.tl0 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.SXT 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.443 19:15:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.443 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.443 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha384.PND ]] 00:19:12.443 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.PND 00:19:12.443 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.443 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.0uM 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha256.D9C ]] 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.D9C 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.gQD 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n '' ]] 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@179 -- # connect_authenticate sha512 ffdhe8192 3 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:12.444 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:13.385 nvme0n1 00:19:13.385 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:19:13.385 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:19:13.385 19:15:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:19:13.645 { 00:19:13.645 "cntlid": 1, 00:19:13.645 "qid": 0, 00:19:13.645 "state": "enabled", 00:19:13.645 "thread": "nvmf_tgt_poll_group_000", 00:19:13.645 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:13.645 "listen_address": { 00:19:13.645 "trtype": "TCP", 00:19:13.645 "adrfam": "IPv4", 00:19:13.645 "traddr": "10.0.0.2", 00:19:13.645 "trsvcid": "4420" 00:19:13.645 }, 00:19:13.645 "peer_address": { 00:19:13.645 "trtype": "TCP", 00:19:13.645 "adrfam": "IPv4", 00:19:13.645 "traddr": "10.0.0.1", 00:19:13.645 "trsvcid": "35610" 00:19:13.645 }, 00:19:13.645 "auth": { 00:19:13.645 "state": "completed", 00:19:13.645 "digest": "sha512", 00:19:13.645 "dhgroup": "ffdhe8192" 00:19:13.645 } 00:19:13.645 } 00:19:13.645 ]' 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:13.645 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:13.906 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:19:13.906 19:15:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:19:14.848 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:19:14.848 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:19:14.848 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:14.848 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:14.848 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:14.848 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@182 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key3 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@183 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@184 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key3 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:14.849 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:15.110 request: 00:19:15.110 { 00:19:15.110 "name": "nvme0", 00:19:15.110 "trtype": "tcp", 00:19:15.110 "traddr": "10.0.0.2", 00:19:15.110 "adrfam": "ipv4", 00:19:15.110 "trsvcid": "4420", 00:19:15.110 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:19:15.110 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:15.110 "prchk_reftag": false, 00:19:15.110 "prchk_guard": false, 00:19:15.110 "hdgst": false, 00:19:15.110 "ddgst": false, 00:19:15.110 "dhchap_key": "key3", 00:19:15.110 "allow_unrecognized_csi": false, 00:19:15.110 "method": "bdev_nvme_attach_controller", 00:19:15.110 "req_id": 1 00:19:15.110 } 00:19:15.110 Got JSON-RPC error response 00:19:15.110 response: 00:19:15.110 { 00:19:15.110 "code": -5, 00:19:15.110 "message": "Input/output error" 00:19:15.110 } 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # IFS=, 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@188 -- # printf %s sha256,sha384,sha512 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # hostrpc bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@193 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key3 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:15.110 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:19:15.371 request: 00:19:15.371 { 00:19:15.371 "name": "nvme0", 00:19:15.371 "trtype": "tcp", 00:19:15.371 "traddr": "10.0.0.2", 00:19:15.371 "adrfam": "ipv4", 00:19:15.371 "trsvcid": "4420", 00:19:15.371 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:19:15.371 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:15.371 "prchk_reftag": false, 00:19:15.371 "prchk_guard": false, 00:19:15.371 "hdgst": false, 00:19:15.371 "ddgst": false, 00:19:15.371 "dhchap_key": "key3", 00:19:15.371 "allow_unrecognized_csi": false, 00:19:15.371 "method": "bdev_nvme_attach_controller", 00:19:15.371 "req_id": 1 00:19:15.371 } 00:19:15.371 Got JSON-RPC error response 00:19:15.371 response: 00:19:15.371 { 00:19:15.371 "code": -5, 00:19:15.371 "message": "Input/output error" 00:19:15.371 } 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s sha256,sha384,sha512 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:19:15.371 19:15:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@208 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@209 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@210 -- # NOT bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:19:15.659 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:19:15.945 request: 00:19:15.945 { 00:19:15.945 "name": "nvme0", 00:19:15.945 "trtype": "tcp", 00:19:15.945 "traddr": "10.0.0.2", 00:19:15.945 "adrfam": "ipv4", 00:19:15.945 "trsvcid": "4420", 00:19:15.945 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:19:15.945 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:15.945 "prchk_reftag": false, 00:19:15.945 "prchk_guard": false, 00:19:15.945 "hdgst": false, 00:19:15.945 "ddgst": false, 00:19:15.945 "dhchap_key": "key0", 00:19:15.945 "dhchap_ctrlr_key": "key1", 00:19:15.945 "allow_unrecognized_csi": false, 00:19:15.945 "method": "bdev_nvme_attach_controller", 00:19:15.945 "req_id": 1 00:19:15.945 } 00:19:15.945 Got JSON-RPC error response 00:19:15.945 response: 00:19:15.945 { 00:19:15.945 "code": -5, 00:19:15.945 "message": "Input/output error" 00:19:15.945 } 00:19:15.945 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:15.945 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:15.945 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:15.945 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:15.945 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@213 -- # bdev_connect -b nvme0 --dhchap-key key0 00:19:15.945 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:19:15.945 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:19:16.212 nvme0n1 00:19:16.212 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # hostrpc bdev_nvme_get_controllers 00:19:16.212 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # jq -r '.[].name' 00:19:16.212 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:16.516 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:16.516 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@215 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:16.516 19:15:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:16.516 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@218 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 00:19:16.516 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:16.516 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:16.516 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:16.516 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@219 -- # bdev_connect -b nvme0 --dhchap-key key1 00:19:16.516 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:19:16.516 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:19:17.457 nvme0n1 00:19:17.457 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # hostrpc bdev_nvme_get_controllers 00:19:17.457 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # jq -r '.[].name' 00:19:17.457 19:15:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@222 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # hostrpc bdev_nvme_get_controllers 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # jq -r '.[].name' 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@225 -- # nvme_connect --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:19:17.718 19:15:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid 00539ede-7deb-ec11-9bc7-a4bf01928396 -l 0 --dhchap-secret DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: --dhchap-ctrl-secret DHHC-1:03:MDA4N2UxNzU0YWZkNThkN2Y0YzYxYzE0OTZlMzg0YTBlNDhmYzRjMmI0NGQzNzkxODg3Y2QxOWI3ZWM1NzJkYlE21QM=: 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nvme_get_ctrlr 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@41 -- # local dev 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@43 -- # for dev in /sys/devices/virtual/nvme-fabrics/ctl/nvme* 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # [[ nqn.2024-03.io.spdk:cnode0 == \n\q\n\.\2\0\2\4\-\0\3\.\i\o\.\s\p\d\k\:\c\n\o\d\e\0 ]] 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # echo nvme0 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # break 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nctrlr=nvme0 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@227 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@228 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key1 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:19:18.657 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:19:19.228 request: 00:19:19.228 { 00:19:19.228 "name": "nvme0", 00:19:19.228 "trtype": "tcp", 00:19:19.228 "traddr": "10.0.0.2", 00:19:19.228 "adrfam": "ipv4", 00:19:19.228 "trsvcid": "4420", 00:19:19.228 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:19:19.228 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396", 00:19:19.228 "prchk_reftag": false, 00:19:19.228 "prchk_guard": false, 00:19:19.228 "hdgst": false, 00:19:19.228 "ddgst": false, 00:19:19.228 "dhchap_key": "key1", 00:19:19.228 "allow_unrecognized_csi": false, 00:19:19.228 "method": "bdev_nvme_attach_controller", 00:19:19.228 "req_id": 1 00:19:19.228 } 00:19:19.228 Got JSON-RPC error response 00:19:19.228 response: 00:19:19.228 { 00:19:19.228 "code": -5, 00:19:19.228 "message": "Input/output error" 00:19:19.228 } 00:19:19.228 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:19.228 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:19.228 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:19.228 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:19.229 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@229 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:19.229 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:19.229 19:15:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:20.168 nvme0n1 00:19:20.168 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # hostrpc bdev_nvme_get_controllers 00:19:20.168 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # jq -r '.[].name' 00:19:20.168 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:20.168 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:20.168 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@231 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:20.168 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:20.428 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@233 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:20.428 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:20.428 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:20.428 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:20.428 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@234 -- # bdev_connect -b nvme0 00:19:20.428 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:19:20.428 19:15:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:19:20.688 nvme0n1 00:19:20.688 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # hostrpc bdev_nvme_get_controllers 00:19:20.688 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # jq -r '.[].name' 00:19:20.688 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@236 -- # hostrpc bdev_nvme_detach_controller nvme0 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@239 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key key3 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@240 -- # nvme_set_keys nvme0 DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: '' 2s 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key=DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey= 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: ]] 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # echo DHHC-1:01:M2NlN2FjNGI2YjBjNzJmY2VlOWRkZDkzZDdiNDQyZjhCbBqy: 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z '' ]] 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:19:20.949 19:15:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@241 -- # waitforblk nvme0n1 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1239 -- # local i=0 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n1 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1250 -- # return 0 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@243 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key1 --dhchap-ctrlr-key key2 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@244 -- # nvme_set_keys nvme0 '' DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: 2s 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key= 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey=DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z '' ]] 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: ]] 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # echo DHHC-1:02:YTU5ZDFhYTQwODAwNDM0OWI0MmEyMzExNTU2YzIxYzVlM2QwYTU1NWQ0NWMwMWY0Kt6qgQ==: 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:19:23.519 19:15:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@245 -- # waitforblk nvme0n1 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1239 -- # local i=0 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n1 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1250 -- # return 0 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@246 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:19:25.431 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@249 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key key1 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@250 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:19:25.431 19:15:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:19:26.003 nvme0n1 00:19:26.003 19:15:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@252 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:26.003 19:15:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:26.003 19:15:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:26.003 19:15:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:26.003 19:15:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@253 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:26.003 19:15:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:26.573 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # hostrpc bdev_nvme_get_controllers 00:19:26.573 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # jq -r '.[].name' 00:19:26.573 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:26.832 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:26.832 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@256 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:26.832 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:26.832 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:26.832 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:26.832 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@257 -- # hostrpc bdev_nvme_set_keys nvme0 00:19:26.832 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # hostrpc bdev_nvme_get_controllers 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # jq -r '.[].name' 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@260 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@261 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=hostrpc 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t hostrpc 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:19:27.093 19:15:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:19:27.664 request: 00:19:27.664 { 00:19:27.664 "name": "nvme0", 00:19:27.664 "dhchap_key": "key1", 00:19:27.664 "dhchap_ctrlr_key": "key3", 00:19:27.664 "method": "bdev_nvme_set_keys", 00:19:27.664 "req_id": 1 00:19:27.664 } 00:19:27.664 Got JSON-RPC error response 00:19:27.664 response: 00:19:27.664 { 00:19:27.664 "code": -13, 00:19:27.664 "message": "Permission denied" 00:19:27.664 } 00:19:27.664 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:27.664 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:27.664 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:27.664 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:27.664 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:19:27.664 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:19:27.664 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:27.924 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 1 != 0 )) 00:19:27.924 19:15:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@263 -- # sleep 1s 00:19:28.865 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:19:28.865 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:19:28.865 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 0 != 0 )) 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@267 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key0 --dhchap-ctrlr-key key1 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@268 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:19:29.125 19:15:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:19:30.068 nvme0n1 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@270 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --dhchap-key key2 --dhchap-ctrlr-key key3 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@271 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=hostrpc 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t hostrpc 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:19:30.068 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:19:30.329 request: 00:19:30.329 { 00:19:30.329 "name": "nvme0", 00:19:30.329 "dhchap_key": "key2", 00:19:30.329 "dhchap_ctrlr_key": "key0", 00:19:30.329 "method": "bdev_nvme_set_keys", 00:19:30.329 "req_id": 1 00:19:30.329 } 00:19:30.329 Got JSON-RPC error response 00:19:30.329 response: 00:19:30.329 { 00:19:30.329 "code": -13, 00:19:30.329 "message": "Permission denied" 00:19:30.329 } 00:19:30.329 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:19:30.329 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:30.329 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:30.329 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:30.329 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:19:30.329 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:19:30.329 19:15:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:30.590 19:15:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 1 != 0 )) 00:19:30.590 19:15:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@273 -- # sleep 1s 00:19:31.531 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:19:31.531 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:19:31.531 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 0 != 0 )) 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@276 -- # trap - SIGINT SIGTERM EXIT 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@277 -- # cleanup 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@21 -- # killprocess 3828591 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # '[' -z 3828591 ']' 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@958 -- # kill -0 3828591 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # uname 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3828591 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3828591' 00:19:31.793 killing process with pid 3828591 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@973 -- # kill 3828591 00:19:31.793 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@978 -- # wait 3828591 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@22 -- # nvmftestfini 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@121 -- # sync 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@124 -- # set +e 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:19:32.053 rmmod nvme_tcp 00:19:32.053 rmmod nvme_fabrics 00:19:32.053 rmmod nvme_keyring 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@128 -- # set -e 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@129 -- # return 0 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@517 -- # '[' -n 3856531 ']' 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@518 -- # killprocess 3856531 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # '[' -z 3856531 ']' 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@958 -- # kill -0 3856531 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # uname 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:32.053 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3856531 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3856531' 00:19:32.314 killing process with pid 3856531 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@973 -- # kill 3856531 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@978 -- # wait 3856531 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@297 -- # iptr 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@791 -- # iptables-save 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@791 -- # iptables-restore 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:32.314 19:15:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@23 -- # rm -f /tmp/spdk.key-null.uaB /tmp/spdk.key-sha256.SXT /tmp/spdk.key-sha384.0uM /tmp/spdk.key-sha512.gQD /tmp/spdk.key-sha512.tl0 /tmp/spdk.key-sha384.PND /tmp/spdk.key-sha256.D9C '' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf-auth.log 00:19:34.859 00:19:34.859 real 2m47.430s 00:19:34.859 user 6m11.600s 00:19:34.859 sys 0m25.546s 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:19:34.859 ************************************ 00:19:34.859 END TEST nvmf_auth_target 00:19:34.859 ************************************ 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@39 -- # '[' tcp = tcp ']' 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@40 -- # run_test nvmf_bdevio_no_huge /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:34.859 ************************************ 00:19:34.859 START TEST nvmf_bdevio_no_huge 00:19:34.859 ************************************ 00:19:34.859 19:15:46 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:19:34.859 * Looking for test storage... 00:19:34.859 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1693 -- # lcov --version 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # IFS=.-: 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # read -ra ver1 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # IFS=.-: 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # read -ra ver2 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@338 -- # local 'op=<' 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@340 -- # ver1_l=2 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@341 -- # ver2_l=1 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@344 -- # case "$op" in 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@345 -- # : 1 00:19:34.859 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # decimal 1 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=1 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 1 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # ver1[v]=1 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # decimal 2 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=2 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 2 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # ver2[v]=2 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # return 0 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:19:34.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:34.860 --rc genhtml_branch_coverage=1 00:19:34.860 --rc genhtml_function_coverage=1 00:19:34.860 --rc genhtml_legend=1 00:19:34.860 --rc geninfo_all_blocks=1 00:19:34.860 --rc geninfo_unexecuted_blocks=1 00:19:34.860 00:19:34.860 ' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:19:34.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:34.860 --rc genhtml_branch_coverage=1 00:19:34.860 --rc genhtml_function_coverage=1 00:19:34.860 --rc genhtml_legend=1 00:19:34.860 --rc geninfo_all_blocks=1 00:19:34.860 --rc geninfo_unexecuted_blocks=1 00:19:34.860 00:19:34.860 ' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:19:34.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:34.860 --rc genhtml_branch_coverage=1 00:19:34.860 --rc genhtml_function_coverage=1 00:19:34.860 --rc genhtml_legend=1 00:19:34.860 --rc geninfo_all_blocks=1 00:19:34.860 --rc geninfo_unexecuted_blocks=1 00:19:34.860 00:19:34.860 ' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:19:34.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:34.860 --rc genhtml_branch_coverage=1 00:19:34.860 --rc genhtml_function_coverage=1 00:19:34.860 --rc genhtml_legend=1 00:19:34.860 --rc geninfo_all_blocks=1 00:19:34.860 --rc geninfo_unexecuted_blocks=1 00:19:34.860 00:19:34.860 ' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # uname -s 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@15 -- # shopt -s extglob 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@5 -- # export PATH 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@51 -- # : 0 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:34.860 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@14 -- # nvmftestinit 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@476 -- # prepare_net_devs 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@438 -- # local -g is_hw=no 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # remove_spdk_ns 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@309 -- # xtrace_disable 00:19:34.860 19:15:47 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # pci_devs=() 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # local -a pci_devs 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # pci_net_devs=() 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # pci_drivers=() 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # local -A pci_drivers 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # net_devs=() 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # local -ga net_devs 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # e810=() 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # local -ga e810 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # x722=() 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # local -ga x722 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # mlx=() 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # local -ga mlx 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:19:42.998 Found 0000:31:00.0 (0x8086 - 0x159b) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:19:42.998 Found 0000:31:00.1 (0x8086 - 0x159b) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:19:42.998 Found net devices under 0000:31:00.0: cvl_0_0 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:19:42.998 Found net devices under 0000:31:00.1: cvl_0_1 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # is_hw=yes 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:42.998 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:19:42.999 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:19:43.259 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:43.259 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.682 ms 00:19:43.259 00:19:43.259 --- 10.0.0.2 ping statistics --- 00:19:43.259 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:43.259 rtt min/avg/max/mdev = 0.682/0.682/0.682/0.000 ms 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:43.259 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:43.259 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.287 ms 00:19:43.259 00:19:43.259 --- 10.0.0.1 ping statistics --- 00:19:43.259 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:43.259 rtt min/avg/max/mdev = 0.287/0.287/0.287/0.000 ms 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@450 -- # return 0 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@726 -- # xtrace_disable 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@509 -- # nvmfpid=3865562 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@510 -- # waitforlisten 3865562 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@835 -- # '[' -z 3865562 ']' 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:43.259 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:43.259 19:15:55 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --no-huge -s 1024 -m 0x78 00:19:43.259 [2024-11-26 19:15:55.761295] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:19:43.259 [2024-11-26 19:15:55.761366] [ DPDK EAL parameters: nvmf -c 0x78 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk0 --proc-type=auto ] 00:19:43.259 [2024-11-26 19:15:55.877081] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:19:43.520 [2024-11-26 19:15:55.937480] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:43.520 [2024-11-26 19:15:55.937530] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:43.520 [2024-11-26 19:15:55.937538] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:43.520 [2024-11-26 19:15:55.937545] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:43.520 [2024-11-26 19:15:55.937551] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:43.520 [2024-11-26 19:15:55.939124] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:19:43.520 [2024-11-26 19:15:55.939281] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:19:43.520 [2024-11-26 19:15:55.939437] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:19:43.520 [2024-11-26 19:15:55.939439] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@868 -- # return 0 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@732 -- # xtrace_disable 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:44.091 [2024-11-26 19:15:56.629417] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:44.091 Malloc0 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:44.091 [2024-11-26 19:15:56.683221] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 --no-huge -s 1024 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@560 -- # config=() 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@560 -- # local subsystem config 00:19:44.091 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:19:44.092 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:19:44.092 { 00:19:44.092 "params": { 00:19:44.092 "name": "Nvme$subsystem", 00:19:44.092 "trtype": "$TEST_TRANSPORT", 00:19:44.092 "traddr": "$NVMF_FIRST_TARGET_IP", 00:19:44.092 "adrfam": "ipv4", 00:19:44.092 "trsvcid": "$NVMF_PORT", 00:19:44.092 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:19:44.092 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:19:44.092 "hdgst": ${hdgst:-false}, 00:19:44.092 "ddgst": ${ddgst:-false} 00:19:44.092 }, 00:19:44.092 "method": "bdev_nvme_attach_controller" 00:19:44.092 } 00:19:44.092 EOF 00:19:44.092 )") 00:19:44.092 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@582 -- # cat 00:19:44.092 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@584 -- # jq . 00:19:44.092 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@585 -- # IFS=, 00:19:44.092 19:15:56 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:19:44.092 "params": { 00:19:44.092 "name": "Nvme1", 00:19:44.092 "trtype": "tcp", 00:19:44.092 "traddr": "10.0.0.2", 00:19:44.092 "adrfam": "ipv4", 00:19:44.092 "trsvcid": "4420", 00:19:44.092 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:44.092 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:44.092 "hdgst": false, 00:19:44.092 "ddgst": false 00:19:44.092 }, 00:19:44.092 "method": "bdev_nvme_attach_controller" 00:19:44.092 }' 00:19:44.352 [2024-11-26 19:15:56.752670] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:19:44.352 [2024-11-26 19:15:56.752743] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk_pid3865907 ] 00:19:44.352 [2024-11-26 19:15:56.843122] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:19:44.352 [2024-11-26 19:15:56.898414] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:44.352 [2024-11-26 19:15:56.898529] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:44.352 [2024-11-26 19:15:56.898532] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:44.612 I/O targets: 00:19:44.612 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:19:44.612 00:19:44.612 00:19:44.612 CUnit - A unit testing framework for C - Version 2.1-3 00:19:44.612 http://cunit.sourceforge.net/ 00:19:44.612 00:19:44.612 00:19:44.612 Suite: bdevio tests on: Nvme1n1 00:19:44.612 Test: blockdev write read block ...passed 00:19:44.612 Test: blockdev write zeroes read block ...passed 00:19:44.612 Test: blockdev write zeroes read no split ...passed 00:19:44.872 Test: blockdev write zeroes read split ...passed 00:19:44.872 Test: blockdev write zeroes read split partial ...passed 00:19:44.872 Test: blockdev reset ...[2024-11-26 19:15:57.281139] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:19:44.872 [2024-11-26 19:15:57.281200] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xa6bf70 (9): Bad file descriptor 00:19:44.872 [2024-11-26 19:15:57.342783] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:19:44.872 passed 00:19:44.872 Test: blockdev write read 8 blocks ...passed 00:19:44.872 Test: blockdev write read size > 128k ...passed 00:19:44.872 Test: blockdev write read invalid size ...passed 00:19:44.872 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:19:44.872 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:19:44.872 Test: blockdev write read max offset ...passed 00:19:44.872 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:19:44.872 Test: blockdev writev readv 8 blocks ...passed 00:19:45.132 Test: blockdev writev readv 30 x 1block ...passed 00:19:45.132 Test: blockdev writev readv block ...passed 00:19:45.132 Test: blockdev writev readv size > 128k ...passed 00:19:45.132 Test: blockdev writev readv size > 128k in two iovs ...passed 00:19:45.132 Test: blockdev comparev and writev ...[2024-11-26 19:15:57.569474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.569499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.569512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.569517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.569963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.569972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.569981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.569987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.570471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.570479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.570490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.570496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.570997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.571005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.571014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:19:45.132 [2024-11-26 19:15:57.571020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:19:45.132 passed 00:19:45.132 Test: blockdev nvme passthru rw ...passed 00:19:45.132 Test: blockdev nvme passthru vendor specific ...[2024-11-26 19:15:57.654752] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:19:45.132 [2024-11-26 19:15:57.654763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.655080] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:19:45.132 [2024-11-26 19:15:57.655088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.655415] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:19:45.132 [2024-11-26 19:15:57.655423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:19:45.132 [2024-11-26 19:15:57.655740] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:19:45.132 [2024-11-26 19:15:57.655748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:19:45.132 passed 00:19:45.132 Test: blockdev nvme admin passthru ...passed 00:19:45.132 Test: blockdev copy ...passed 00:19:45.132 00:19:45.133 Run Summary: Type Total Ran Passed Failed Inactive 00:19:45.133 suites 1 1 n/a 0 0 00:19:45.133 tests 23 23 23 0 0 00:19:45.133 asserts 152 152 152 0 n/a 00:19:45.133 00:19:45.133 Elapsed time = 1.235 seconds 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@30 -- # nvmftestfini 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@516 -- # nvmfcleanup 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@121 -- # sync 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@124 -- # set +e 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@125 -- # for i in {1..20} 00:19:45.393 19:15:57 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:19:45.393 rmmod nvme_tcp 00:19:45.393 rmmod nvme_fabrics 00:19:45.653 rmmod nvme_keyring 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@128 -- # set -e 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@129 -- # return 0 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@517 -- # '[' -n 3865562 ']' 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@518 -- # killprocess 3865562 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@954 -- # '[' -z 3865562 ']' 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@958 -- # kill -0 3865562 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@959 -- # uname 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3865562 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@960 -- # process_name=reactor_3 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@964 -- # '[' reactor_3 = sudo ']' 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3865562' 00:19:45.653 killing process with pid 3865562 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@973 -- # kill 3865562 00:19:45.653 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@978 -- # wait 3865562 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@297 -- # iptr 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@791 -- # iptables-save 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@791 -- # iptables-restore 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@302 -- # remove_spdk_ns 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:45.914 19:15:58 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:19:48.458 00:19:48.458 real 0m13.604s 00:19:48.458 user 0m14.430s 00:19:48.458 sys 0m7.501s 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1130 -- # xtrace_disable 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:19:48.458 ************************************ 00:19:48.458 END TEST nvmf_bdevio_no_huge 00:19:48.458 ************************************ 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@41 -- # run_test nvmf_tls /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:48.458 ************************************ 00:19:48.458 START TEST nvmf_tls 00:19:48.458 ************************************ 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:19:48.458 * Looking for test storage... 00:19:48.458 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1693 -- # lcov --version 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # IFS=.-: 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # read -ra ver1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # IFS=.-: 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # read -ra ver2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@338 -- # local 'op=<' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@340 -- # ver1_l=2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@341 -- # ver2_l=1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@344 -- # case "$op" in 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@345 -- # : 1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # decimal 1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # ver1[v]=1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # decimal 2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # ver2[v]=2 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # return 0 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:19:48.458 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:48.458 --rc genhtml_branch_coverage=1 00:19:48.458 --rc genhtml_function_coverage=1 00:19:48.458 --rc genhtml_legend=1 00:19:48.458 --rc geninfo_all_blocks=1 00:19:48.458 --rc geninfo_unexecuted_blocks=1 00:19:48.458 00:19:48.458 ' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:19:48.458 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:48.458 --rc genhtml_branch_coverage=1 00:19:48.458 --rc genhtml_function_coverage=1 00:19:48.458 --rc genhtml_legend=1 00:19:48.458 --rc geninfo_all_blocks=1 00:19:48.458 --rc geninfo_unexecuted_blocks=1 00:19:48.458 00:19:48.458 ' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:19:48.458 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:48.458 --rc genhtml_branch_coverage=1 00:19:48.458 --rc genhtml_function_coverage=1 00:19:48.458 --rc genhtml_legend=1 00:19:48.458 --rc geninfo_all_blocks=1 00:19:48.458 --rc geninfo_unexecuted_blocks=1 00:19:48.458 00:19:48.458 ' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:19:48.458 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:48.458 --rc genhtml_branch_coverage=1 00:19:48.458 --rc genhtml_function_coverage=1 00:19:48.458 --rc genhtml_legend=1 00:19:48.458 --rc geninfo_all_blocks=1 00:19:48.458 --rc geninfo_unexecuted_blocks=1 00:19:48.458 00:19:48.458 ' 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # uname -s 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:19:48.458 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@15 -- # shopt -s extglob 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@5 -- # export PATH 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@51 -- # : 0 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:48.459 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@63 -- # nvmftestinit 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@476 -- # prepare_net_devs 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@438 -- # local -g is_hw=no 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # remove_spdk_ns 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@309 -- # xtrace_disable 00:19:48.459 19:16:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # pci_devs=() 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # local -a pci_devs 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # pci_net_devs=() 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # pci_drivers=() 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # local -A pci_drivers 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # net_devs=() 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # local -ga net_devs 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # e810=() 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # local -ga e810 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # x722=() 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # local -ga x722 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # mlx=() 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # local -ga mlx 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:19:56.597 Found 0000:31:00.0 (0x8086 - 0x159b) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:19:56.597 Found 0000:31:00.1 (0x8086 - 0x159b) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:19:56.597 Found net devices under 0000:31:00.0: cvl_0_0 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:19:56.597 Found net devices under 0000:31:00.1: cvl_0_1 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # is_hw=yes 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:19:56.597 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:19:56.859 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:56.859 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.676 ms 00:19:56.859 00:19:56.859 --- 10.0.0.2 ping statistics --- 00:19:56.859 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:56.859 rtt min/avg/max/mdev = 0.676/0.676/0.676/0.000 ms 00:19:56.859 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:57.121 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:57.121 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.282 ms 00:19:57.121 00:19:57.121 --- 10.0.0.1 ping statistics --- 00:19:57.121 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:57.121 rtt min/avg/max/mdev = 0.282/0.282/0.282/0.000 ms 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@450 -- # return 0 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@64 -- # nvmfappstart -m 0x2 --wait-for-rpc 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3870937 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3870937 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 --wait-for-rpc 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3870937 ']' 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:57.121 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:57.121 19:16:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:57.121 [2024-11-26 19:16:09.593214] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:19:57.121 [2024-11-26 19:16:09.593279] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:57.121 [2024-11-26 19:16:09.701874] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:57.383 [2024-11-26 19:16:09.751462] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:57.383 [2024-11-26 19:16:09.751513] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:57.383 [2024-11-26 19:16:09.751522] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:57.383 [2024-11-26 19:16:09.751529] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:57.383 [2024-11-26 19:16:09.751536] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:57.383 [2024-11-26 19:16:09.752324] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@66 -- # '[' tcp '!=' tcp ']' 00:19:57.954 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_set_default_impl -i ssl 00:19:58.214 true 00:19:58.214 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:19:58.214 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # jq -r .tls_version 00:19:58.214 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # version=0 00:19:58.214 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@75 -- # [[ 0 != \0 ]] 00:19:58.214 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:19:58.474 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:19:58.474 19:16:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # jq -r .tls_version 00:19:58.734 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # version=13 00:19:58.734 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@83 -- # [[ 13 != \1\3 ]] 00:19:58.734 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 7 00:19:58.734 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:19:58.734 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # jq -r .tls_version 00:19:58.995 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # version=7 00:19:58.995 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@91 -- # [[ 7 != \7 ]] 00:19:58.995 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:19:58.995 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # jq -r .enable_ktls 00:19:59.256 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # ktls=false 00:19:59.256 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@98 -- # [[ false != \f\a\l\s\e ]] 00:19:59.256 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@104 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --enable-ktls 00:19:59.522 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:19:59.522 19:16:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # jq -r .enable_ktls 00:19:59.522 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # ktls=true 00:19:59.522 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@106 -- # [[ true != \t\r\u\e ]] 00:19:59.522 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@112 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --disable-ktls 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # jq -r .enable_ktls 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # ktls=false 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@114 -- # [[ false != \f\a\l\s\e ]] 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # format_interchange_psk 00112233445566778899aabbccddeeff 1 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 1 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # local prefix key digest 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # digest=1 00:19:59.783 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@733 -- # python - 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # key=NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # format_interchange_psk ffeeddccbbaa99887766554433221100 1 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 ffeeddccbbaa99887766554433221100 1 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # local prefix key digest 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # key=ffeeddccbbaa99887766554433221100 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # digest=1 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@733 -- # python - 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # key_2=NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # mktemp 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # key_path=/tmp/tmp.C8WQVOIq90 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # mktemp 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # key_2_path=/tmp/tmp.KAnagrqvyX 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@125 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@126 -- # echo -n NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@128 -- # chmod 0600 /tmp/tmp.C8WQVOIq90 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@129 -- # chmod 0600 /tmp/tmp.KAnagrqvyX 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:20:00.044 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@132 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_start_init 00:20:00.305 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@134 -- # setup_nvmf_tgt /tmp/tmp.C8WQVOIq90 00:20:00.305 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.C8WQVOIq90 00:20:00.305 19:16:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:20:00.566 [2024-11-26 19:16:13.047441] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:00.566 19:16:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:20:00.827 19:16:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:20:00.827 [2024-11-26 19:16:13.368217] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:20:00.827 [2024-11-26 19:16:13.368405] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:00.827 19:16:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:20:01.087 malloc0 00:20:01.087 19:16:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:20:01.087 19:16:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.C8WQVOIq90 00:20:01.347 19:16:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:20:01.607 19:16:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@138 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -S ssl -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 hostnqn:nqn.2016-06.io.spdk:host1' --psk-path /tmp/tmp.C8WQVOIq90 00:20:11.601 Initializing NVMe Controllers 00:20:11.601 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:20:11.601 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:20:11.601 Initialization complete. Launching workers. 00:20:11.601 ======================================================== 00:20:11.602 Latency(us) 00:20:11.602 Device Information : IOPS MiB/s Average min max 00:20:11.602 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 18688.33 73.00 3424.57 1152.18 4234.32 00:20:11.602 ======================================================== 00:20:11.602 Total : 18688.33 73.00 3424.57 1152.18 4234.32 00:20:11.602 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@144 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.C8WQVOIq90 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.C8WQVOIq90 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=3873849 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 3873849 /var/tmp/bdevperf.sock 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3873849 ']' 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:11.602 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:11.602 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:11.602 [2024-11-26 19:16:24.178728] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:11.602 [2024-11-26 19:16:24.178782] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3873849 ] 00:20:11.863 [2024-11-26 19:16:24.241869] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:11.863 [2024-11-26 19:16:24.270831] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:11.863 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:11.863 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:11.863 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.C8WQVOIq90 00:20:12.123 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:12.123 [2024-11-26 19:16:24.669162] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:12.123 TLSTESTn1 00:20:12.384 19:16:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:20:12.384 Running I/O for 10 seconds... 00:20:14.322 6023.00 IOPS, 23.53 MiB/s [2024-11-26T18:16:27.925Z] 6268.00 IOPS, 24.48 MiB/s [2024-11-26T18:16:28.947Z] 6289.00 IOPS, 24.57 MiB/s [2024-11-26T18:16:29.888Z] 6329.00 IOPS, 24.72 MiB/s [2024-11-26T18:16:31.271Z] 6132.60 IOPS, 23.96 MiB/s [2024-11-26T18:16:32.209Z] 6190.17 IOPS, 24.18 MiB/s [2024-11-26T18:16:33.152Z] 6113.86 IOPS, 23.88 MiB/s [2024-11-26T18:16:34.090Z] 6143.75 IOPS, 24.00 MiB/s [2024-11-26T18:16:35.032Z] 6131.00 IOPS, 23.95 MiB/s [2024-11-26T18:16:35.032Z] 6055.90 IOPS, 23.66 MiB/s 00:20:22.407 Latency(us) 00:20:22.407 [2024-11-26T18:16:35.032Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:22.407 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:20:22.407 Verification LBA range: start 0x0 length 0x2000 00:20:22.407 TLSTESTn1 : 10.02 6054.87 23.65 0.00 0.00 21103.85 4696.75 31020.37 00:20:22.407 [2024-11-26T18:16:35.032Z] =================================================================================================================== 00:20:22.407 [2024-11-26T18:16:35.032Z] Total : 6054.87 23.65 0.00 0.00 21103.85 4696.75 31020.37 00:20:22.407 { 00:20:22.407 "results": [ 00:20:22.407 { 00:20:22.407 "job": "TLSTESTn1", 00:20:22.407 "core_mask": "0x4", 00:20:22.407 "workload": "verify", 00:20:22.407 "status": "finished", 00:20:22.407 "verify_range": { 00:20:22.407 "start": 0, 00:20:22.407 "length": 8192 00:20:22.407 }, 00:20:22.407 "queue_depth": 128, 00:20:22.407 "io_size": 4096, 00:20:22.407 "runtime": 10.022849, 00:20:22.407 "iops": 6054.865238416742, 00:20:22.407 "mibps": 23.651817337565397, 00:20:22.407 "io_failed": 0, 00:20:22.407 "io_timeout": 0, 00:20:22.407 "avg_latency_us": 21103.84824075447, 00:20:22.407 "min_latency_us": 4696.746666666667, 00:20:22.407 "max_latency_us": 31020.373333333333 00:20:22.407 } 00:20:22.407 ], 00:20:22.407 "core_count": 1 00:20:22.407 } 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 3873849 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3873849 ']' 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3873849 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3873849 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3873849' 00:20:22.407 killing process with pid 3873849 00:20:22.407 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3873849 00:20:22.407 Received shutdown signal, test time was about 10.000000 seconds 00:20:22.407 00:20:22.407 Latency(us) 00:20:22.407 [2024-11-26T18:16:35.032Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:22.407 [2024-11-26T18:16:35.033Z] =================================================================================================================== 00:20:22.408 [2024-11-26T18:16:35.033Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:20:22.408 19:16:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3873849 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@147 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.KAnagrqvyX 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.KAnagrqvyX 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.KAnagrqvyX 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.KAnagrqvyX 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=3876337 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 3876337 /var/tmp/bdevperf.sock 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3876337 ']' 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:22.668 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:22.668 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:22.668 [2024-11-26 19:16:35.142331] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:22.668 [2024-11-26 19:16:35.142382] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3876337 ] 00:20:22.668 [2024-11-26 19:16:35.207189] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:22.668 [2024-11-26 19:16:35.235053] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:22.928 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:22.928 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:22.928 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.KAnagrqvyX 00:20:22.928 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:23.188 [2024-11-26 19:16:35.657599] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:23.188 [2024-11-26 19:16:35.662099] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:20:23.188 [2024-11-26 19:16:35.662721] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1078990 (107): Transport endpoint is not connected 00:20:23.188 [2024-11-26 19:16:35.663716] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1078990 (9): Bad file descriptor 00:20:23.188 [2024-11-26 19:16:35.664718] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] Ctrlr is in error state 00:20:23.188 [2024-11-26 19:16:35.664726] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:20:23.188 [2024-11-26 19:16:35.664732] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:20:23.188 [2024-11-26 19:16:35.664738] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] in failed state. 00:20:23.188 request: 00:20:23.188 { 00:20:23.188 "name": "TLSTEST", 00:20:23.188 "trtype": "tcp", 00:20:23.188 "traddr": "10.0.0.2", 00:20:23.188 "adrfam": "ipv4", 00:20:23.188 "trsvcid": "4420", 00:20:23.188 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:23.188 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:23.188 "prchk_reftag": false, 00:20:23.188 "prchk_guard": false, 00:20:23.188 "hdgst": false, 00:20:23.188 "ddgst": false, 00:20:23.188 "psk": "key0", 00:20:23.188 "allow_unrecognized_csi": false, 00:20:23.188 "method": "bdev_nvme_attach_controller", 00:20:23.188 "req_id": 1 00:20:23.188 } 00:20:23.188 Got JSON-RPC error response 00:20:23.188 response: 00:20:23.188 { 00:20:23.188 "code": -5, 00:20:23.188 "message": "Input/output error" 00:20:23.188 } 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 3876337 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3876337 ']' 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3876337 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3876337 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3876337' 00:20:23.188 killing process with pid 3876337 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3876337 00:20:23.188 Received shutdown signal, test time was about 10.000000 seconds 00:20:23.188 00:20:23.188 Latency(us) 00:20:23.188 [2024-11-26T18:16:35.813Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:23.188 [2024-11-26T18:16:35.813Z] =================================================================================================================== 00:20:23.188 [2024-11-26T18:16:35.813Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:20:23.188 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3876337 00:20:23.448 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:20:23.448 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:20:23.448 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:23.448 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:23.448 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:23.448 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@150 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.C8WQVOIq90 00:20:23.448 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.C8WQVOIq90 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.C8WQVOIq90 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host2 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.C8WQVOIq90 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=3876388 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 3876388 /var/tmp/bdevperf.sock 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3876388 ']' 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:23.449 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:23.449 19:16:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:23.449 [2024-11-26 19:16:35.905947] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:23.449 [2024-11-26 19:16:35.905999] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3876388 ] 00:20:23.449 [2024-11-26 19:16:35.970484] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:23.449 [2024-11-26 19:16:35.998312] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:23.708 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:23.708 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:23.708 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.C8WQVOIq90 00:20:23.708 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 --psk key0 00:20:23.968 [2024-11-26 19:16:36.416696] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:23.968 [2024-11-26 19:16:36.421967] tcp.c: 969:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:20:23.968 [2024-11-26 19:16:36.421986] posix.c: 573:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:20:23.968 [2024-11-26 19:16:36.422007] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:20:23.968 [2024-11-26 19:16:36.422847] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12a8990 (107): Transport endpoint is not connected 00:20:23.968 [2024-11-26 19:16:36.423842] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12a8990 (9): Bad file descriptor 00:20:23.968 [2024-11-26 19:16:36.424845] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] Ctrlr is in error state 00:20:23.968 [2024-11-26 19:16:36.424852] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:20:23.968 [2024-11-26 19:16:36.424858] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:20:23.968 [2024-11-26 19:16:36.424868] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] in failed state. 00:20:23.968 request: 00:20:23.968 { 00:20:23.968 "name": "TLSTEST", 00:20:23.968 "trtype": "tcp", 00:20:23.968 "traddr": "10.0.0.2", 00:20:23.968 "adrfam": "ipv4", 00:20:23.968 "trsvcid": "4420", 00:20:23.968 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:23.968 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:20:23.968 "prchk_reftag": false, 00:20:23.968 "prchk_guard": false, 00:20:23.968 "hdgst": false, 00:20:23.968 "ddgst": false, 00:20:23.968 "psk": "key0", 00:20:23.968 "allow_unrecognized_csi": false, 00:20:23.968 "method": "bdev_nvme_attach_controller", 00:20:23.968 "req_id": 1 00:20:23.968 } 00:20:23.968 Got JSON-RPC error response 00:20:23.968 response: 00:20:23.968 { 00:20:23.968 "code": -5, 00:20:23.968 "message": "Input/output error" 00:20:23.968 } 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 3876388 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3876388 ']' 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3876388 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3876388 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3876388' 00:20:23.968 killing process with pid 3876388 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3876388 00:20:23.968 Received shutdown signal, test time was about 10.000000 seconds 00:20:23.968 00:20:23.968 Latency(us) 00:20:23.968 [2024-11-26T18:16:36.593Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:23.968 [2024-11-26T18:16:36.593Z] =================================================================================================================== 00:20:23.968 [2024-11-26T18:16:36.593Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:20:23.968 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3876388 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@153 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.C8WQVOIq90 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.C8WQVOIq90 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.C8WQVOIq90 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode2 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.C8WQVOIq90 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=3876685 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 3876685 /var/tmp/bdevperf.sock 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3876685 ']' 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:24.229 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:24.229 [2024-11-26 19:16:36.672031] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:24.229 [2024-11-26 19:16:36.672084] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3876685 ] 00:20:24.229 [2024-11-26 19:16:36.737152] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:24.229 [2024-11-26 19:16:36.765171] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:24.229 19:16:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.C8WQVOIq90 00:20:24.489 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:24.750 [2024-11-26 19:16:37.167371] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:24.750 [2024-11-26 19:16:37.173794] tcp.c: 969:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:20:24.750 [2024-11-26 19:16:37.173811] posix.c: 573:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:20:24.750 [2024-11-26 19:16:37.173829] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:20:24.750 [2024-11-26 19:16:37.174472] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xef3990 (107): Transport endpoint is not connected 00:20:24.750 [2024-11-26 19:16:37.175468] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xef3990 (9): Bad file descriptor 00:20:24.750 [2024-11-26 19:16:37.176470] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 0] Ctrlr is in error state 00:20:24.750 [2024-11-26 19:16:37.176478] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:20:24.750 [2024-11-26 19:16:37.176484] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode2, Operation not permitted 00:20:24.750 [2024-11-26 19:16:37.176490] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 0] in failed state. 00:20:24.750 request: 00:20:24.750 { 00:20:24.750 "name": "TLSTEST", 00:20:24.750 "trtype": "tcp", 00:20:24.750 "traddr": "10.0.0.2", 00:20:24.750 "adrfam": "ipv4", 00:20:24.750 "trsvcid": "4420", 00:20:24.750 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:20:24.750 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:24.750 "prchk_reftag": false, 00:20:24.750 "prchk_guard": false, 00:20:24.750 "hdgst": false, 00:20:24.750 "ddgst": false, 00:20:24.750 "psk": "key0", 00:20:24.750 "allow_unrecognized_csi": false, 00:20:24.750 "method": "bdev_nvme_attach_controller", 00:20:24.750 "req_id": 1 00:20:24.750 } 00:20:24.750 Got JSON-RPC error response 00:20:24.750 response: 00:20:24.750 { 00:20:24.750 "code": -5, 00:20:24.750 "message": "Input/output error" 00:20:24.750 } 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 3876685 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3876685 ']' 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3876685 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3876685 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3876685' 00:20:24.750 killing process with pid 3876685 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3876685 00:20:24.750 Received shutdown signal, test time was about 10.000000 seconds 00:20:24.750 00:20:24.750 Latency(us) 00:20:24.750 [2024-11-26T18:16:37.375Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:24.750 [2024-11-26T18:16:37.375Z] =================================================================================================================== 00:20:24.750 [2024-11-26T18:16:37.375Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3876685 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@156 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk= 00:20:24.750 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=3876761 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 3876761 /var/tmp/bdevperf.sock 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3876761 ']' 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:25.010 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:25.010 [2024-11-26 19:16:37.432761] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:25.010 [2024-11-26 19:16:37.432809] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3876761 ] 00:20:25.010 [2024-11-26 19:16:37.497516] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:25.010 [2024-11-26 19:16:37.525203] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:25.010 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 '' 00:20:25.270 [2024-11-26 19:16:37.762958] keyring.c: 24:keyring_file_check_path: *ERROR*: Non-absolute paths are not allowed: 00:20:25.270 [2024-11-26 19:16:37.762981] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:20:25.270 request: 00:20:25.270 { 00:20:25.270 "name": "key0", 00:20:25.270 "path": "", 00:20:25.270 "method": "keyring_file_add_key", 00:20:25.270 "req_id": 1 00:20:25.270 } 00:20:25.270 Got JSON-RPC error response 00:20:25.270 response: 00:20:25.270 { 00:20:25.270 "code": -1, 00:20:25.270 "message": "Operation not permitted" 00:20:25.270 } 00:20:25.270 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:25.529 [2024-11-26 19:16:37.943491] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:25.529 [2024-11-26 19:16:37.943515] bdev_nvme.c:6722:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:20:25.529 request: 00:20:25.529 { 00:20:25.529 "name": "TLSTEST", 00:20:25.529 "trtype": "tcp", 00:20:25.529 "traddr": "10.0.0.2", 00:20:25.529 "adrfam": "ipv4", 00:20:25.529 "trsvcid": "4420", 00:20:25.529 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:25.529 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:25.529 "prchk_reftag": false, 00:20:25.529 "prchk_guard": false, 00:20:25.529 "hdgst": false, 00:20:25.529 "ddgst": false, 00:20:25.529 "psk": "key0", 00:20:25.529 "allow_unrecognized_csi": false, 00:20:25.529 "method": "bdev_nvme_attach_controller", 00:20:25.529 "req_id": 1 00:20:25.529 } 00:20:25.529 Got JSON-RPC error response 00:20:25.529 response: 00:20:25.529 { 00:20:25.529 "code": -126, 00:20:25.529 "message": "Required key not available" 00:20:25.529 } 00:20:25.529 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 3876761 00:20:25.529 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3876761 ']' 00:20:25.529 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3876761 00:20:25.529 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:25.529 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:25.529 19:16:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3876761 00:20:25.529 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:25.529 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:25.529 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3876761' 00:20:25.529 killing process with pid 3876761 00:20:25.529 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3876761 00:20:25.529 Received shutdown signal, test time was about 10.000000 seconds 00:20:25.529 00:20:25.529 Latency(us) 00:20:25.529 [2024-11-26T18:16:38.154Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:25.529 [2024-11-26T18:16:38.154Z] =================================================================================================================== 00:20:25.529 [2024-11-26T18:16:38.155Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3876761 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@159 -- # killprocess 3870937 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3870937 ']' 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3870937 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:25.530 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3870937 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3870937' 00:20:25.789 killing process with pid 3870937 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3870937 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3870937 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # format_interchange_psk 00112233445566778899aabbccddeeff0011223344556677 2 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff0011223344556677 2 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # local prefix key digest 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff0011223344556677 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # digest=2 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@733 -- # python - 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # key_long=NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # mktemp 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # key_long_path=/tmp/tmp.wq2lUTRnNY 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@162 -- # echo -n NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@163 -- # chmod 0600 /tmp/tmp.wq2lUTRnNY 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@164 -- # nvmfappstart -m 0x2 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3877130 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3877130 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3877130 ']' 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:25.789 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:25.789 19:16:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:26.048 [2024-11-26 19:16:38.446535] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:26.048 [2024-11-26 19:16:38.446619] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:26.048 [2024-11-26 19:16:38.545376] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:26.048 [2024-11-26 19:16:38.575583] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:26.048 [2024-11-26 19:16:38.575613] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:26.048 [2024-11-26 19:16:38.575619] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:26.048 [2024-11-26 19:16:38.575624] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:26.048 [2024-11-26 19:16:38.575628] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:26.048 [2024-11-26 19:16:38.576129] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:26.618 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:26.618 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:26.618 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:26.618 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:26.618 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:26.878 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:26.878 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@166 -- # setup_nvmf_tgt /tmp/tmp.wq2lUTRnNY 00:20:26.878 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wq2lUTRnNY 00:20:26.878 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:20:26.878 [2024-11-26 19:16:39.408661] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:26.878 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:20:27.138 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:20:27.138 [2024-11-26 19:16:39.745498] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:20:27.138 [2024-11-26 19:16:39.745664] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:27.397 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:20:27.397 malloc0 00:20:27.397 19:16:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:20:27.656 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:20:27.656 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@168 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wq2lUTRnNY 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.wq2lUTRnNY 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=3877596 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 3877596 /var/tmp/bdevperf.sock 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3877596 ']' 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:27.915 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:27.915 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:27.915 [2024-11-26 19:16:40.451658] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:27.915 [2024-11-26 19:16:40.451708] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3877596 ] 00:20:27.915 [2024-11-26 19:16:40.515977] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:28.174 [2024-11-26 19:16:40.544777] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:28.174 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:28.174 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:28.174 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:20:28.435 19:16:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:28.435 [2024-11-26 19:16:40.955114] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:28.435 TLSTESTn1 00:20:28.435 19:16:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:20:28.696 Running I/O for 10 seconds... 00:20:30.575 5469.00 IOPS, 21.36 MiB/s [2024-11-26T18:16:44.143Z] 5407.50 IOPS, 21.12 MiB/s [2024-11-26T18:16:45.525Z] 5435.67 IOPS, 21.23 MiB/s [2024-11-26T18:16:46.465Z] 5729.00 IOPS, 22.38 MiB/s [2024-11-26T18:16:47.403Z] 5583.40 IOPS, 21.81 MiB/s [2024-11-26T18:16:48.345Z] 5553.33 IOPS, 21.69 MiB/s [2024-11-26T18:16:49.282Z] 5519.00 IOPS, 21.56 MiB/s [2024-11-26T18:16:50.220Z] 5607.50 IOPS, 21.90 MiB/s [2024-11-26T18:16:51.159Z] 5571.44 IOPS, 21.76 MiB/s [2024-11-26T18:16:51.419Z] 5582.30 IOPS, 21.81 MiB/s 00:20:38.794 Latency(us) 00:20:38.794 [2024-11-26T18:16:51.419Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:38.794 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:20:38.795 Verification LBA range: start 0x0 length 0x2000 00:20:38.795 TLSTESTn1 : 10.01 5586.91 21.82 0.00 0.00 22878.73 5625.17 23702.19 00:20:38.795 [2024-11-26T18:16:51.420Z] =================================================================================================================== 00:20:38.795 [2024-11-26T18:16:51.420Z] Total : 5586.91 21.82 0.00 0.00 22878.73 5625.17 23702.19 00:20:38.795 { 00:20:38.795 "results": [ 00:20:38.795 { 00:20:38.795 "job": "TLSTESTn1", 00:20:38.795 "core_mask": "0x4", 00:20:38.795 "workload": "verify", 00:20:38.795 "status": "finished", 00:20:38.795 "verify_range": { 00:20:38.795 "start": 0, 00:20:38.795 "length": 8192 00:20:38.795 }, 00:20:38.795 "queue_depth": 128, 00:20:38.795 "io_size": 4096, 00:20:38.795 "runtime": 10.014306, 00:20:38.795 "iops": 5586.907370316026, 00:20:38.795 "mibps": 21.823856915296975, 00:20:38.795 "io_failed": 0, 00:20:38.795 "io_timeout": 0, 00:20:38.795 "avg_latency_us": 22878.72636007793, 00:20:38.795 "min_latency_us": 5625.173333333333, 00:20:38.795 "max_latency_us": 23702.18666666667 00:20:38.795 } 00:20:38.795 ], 00:20:38.795 "core_count": 1 00:20:38.795 } 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 3877596 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3877596 ']' 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3877596 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3877596 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3877596' 00:20:38.795 killing process with pid 3877596 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3877596 00:20:38.795 Received shutdown signal, test time was about 10.000000 seconds 00:20:38.795 00:20:38.795 Latency(us) 00:20:38.795 [2024-11-26T18:16:51.420Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:38.795 [2024-11-26T18:16:51.420Z] =================================================================================================================== 00:20:38.795 [2024-11-26T18:16:51.420Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3877596 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@171 -- # chmod 0666 /tmp/tmp.wq2lUTRnNY 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@172 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wq2lUTRnNY 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wq2lUTRnNY 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wq2lUTRnNY 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.wq2lUTRnNY 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=3879854 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 3879854 /var/tmp/bdevperf.sock 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3879854 ']' 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:38.795 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:38.795 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:39.055 [2024-11-26 19:16:51.424367] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:39.055 [2024-11-26 19:16:51.424419] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3879854 ] 00:20:39.055 [2024-11-26 19:16:51.489245] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:39.055 [2024-11-26 19:16:51.516151] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:39.055 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:39.055 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:39.055 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:20:39.315 [2024-11-26 19:16:51.750215] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.wq2lUTRnNY': 0100666 00:20:39.315 [2024-11-26 19:16:51.750244] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:20:39.315 request: 00:20:39.315 { 00:20:39.315 "name": "key0", 00:20:39.315 "path": "/tmp/tmp.wq2lUTRnNY", 00:20:39.315 "method": "keyring_file_add_key", 00:20:39.315 "req_id": 1 00:20:39.315 } 00:20:39.315 Got JSON-RPC error response 00:20:39.315 response: 00:20:39.315 { 00:20:39.315 "code": -1, 00:20:39.315 "message": "Operation not permitted" 00:20:39.315 } 00:20:39.315 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:39.315 [2024-11-26 19:16:51.934759] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:39.315 [2024-11-26 19:16:51.934783] bdev_nvme.c:6722:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:20:39.575 request: 00:20:39.575 { 00:20:39.575 "name": "TLSTEST", 00:20:39.575 "trtype": "tcp", 00:20:39.575 "traddr": "10.0.0.2", 00:20:39.575 "adrfam": "ipv4", 00:20:39.575 "trsvcid": "4420", 00:20:39.575 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:39.575 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:39.575 "prchk_reftag": false, 00:20:39.575 "prchk_guard": false, 00:20:39.575 "hdgst": false, 00:20:39.575 "ddgst": false, 00:20:39.575 "psk": "key0", 00:20:39.575 "allow_unrecognized_csi": false, 00:20:39.575 "method": "bdev_nvme_attach_controller", 00:20:39.575 "req_id": 1 00:20:39.575 } 00:20:39.575 Got JSON-RPC error response 00:20:39.575 response: 00:20:39.575 { 00:20:39.575 "code": -126, 00:20:39.575 "message": "Required key not available" 00:20:39.575 } 00:20:39.575 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 3879854 00:20:39.575 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3879854 ']' 00:20:39.575 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3879854 00:20:39.575 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:39.575 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:39.575 19:16:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3879854 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3879854' 00:20:39.575 killing process with pid 3879854 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3879854 00:20:39.575 Received shutdown signal, test time was about 10.000000 seconds 00:20:39.575 00:20:39.575 Latency(us) 00:20:39.575 [2024-11-26T18:16:52.200Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:39.575 [2024-11-26T18:16:52.200Z] =================================================================================================================== 00:20:39.575 [2024-11-26T18:16:52.200Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3879854 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@175 -- # killprocess 3877130 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3877130 ']' 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3877130 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3877130 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3877130' 00:20:39.575 killing process with pid 3877130 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3877130 00:20:39.575 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3877130 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@176 -- # nvmfappstart -m 0x2 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3880152 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3880152 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3880152 ']' 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:39.834 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:39.834 19:16:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:39.834 [2024-11-26 19:16:52.360931] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:39.834 [2024-11-26 19:16:52.360985] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:40.094 [2024-11-26 19:16:52.458200] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:40.094 [2024-11-26 19:16:52.487044] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:40.094 [2024-11-26 19:16:52.487075] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:40.094 [2024-11-26 19:16:52.487080] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:40.094 [2024-11-26 19:16:52.487085] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:40.094 [2024-11-26 19:16:52.487089] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:40.094 [2024-11-26 19:16:52.487586] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@178 -- # NOT setup_nvmf_tgt /tmp/tmp.wq2lUTRnNY 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg setup_nvmf_tgt /tmp/tmp.wq2lUTRnNY 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=setup_nvmf_tgt 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t setup_nvmf_tgt 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # setup_nvmf_tgt /tmp/tmp.wq2lUTRnNY 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wq2lUTRnNY 00:20:40.665 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:20:40.924 [2024-11-26 19:16:53.352412] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:40.924 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:20:40.924 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:20:41.204 [2024-11-26 19:16:53.689255] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:20:41.204 [2024-11-26 19:16:53.689447] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:41.204 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:20:41.464 malloc0 00:20:41.464 19:16:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:20:41.464 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:20:41.723 [2024-11-26 19:16:54.180397] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.wq2lUTRnNY': 0100666 00:20:41.723 [2024-11-26 19:16:54.180422] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:20:41.723 request: 00:20:41.723 { 00:20:41.723 "name": "key0", 00:20:41.723 "path": "/tmp/tmp.wq2lUTRnNY", 00:20:41.723 "method": "keyring_file_add_key", 00:20:41.723 "req_id": 1 00:20:41.723 } 00:20:41.723 Got JSON-RPC error response 00:20:41.723 response: 00:20:41.723 { 00:20:41.723 "code": -1, 00:20:41.723 "message": "Operation not permitted" 00:20:41.723 } 00:20:41.723 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:20:41.723 [2024-11-26 19:16:54.332794] tcp.c:3792:nvmf_tcp_subsystem_add_host: *ERROR*: Key 'key0' does not exist 00:20:41.723 [2024-11-26 19:16:54.332821] subsystem.c:1051:spdk_nvmf_subsystem_add_host_ext: *ERROR*: Unable to add host to TCP transport 00:20:41.723 request: 00:20:41.723 { 00:20:41.723 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:41.723 "host": "nqn.2016-06.io.spdk:host1", 00:20:41.723 "psk": "key0", 00:20:41.723 "method": "nvmf_subsystem_add_host", 00:20:41.723 "req_id": 1 00:20:41.723 } 00:20:41.723 Got JSON-RPC error response 00:20:41.723 response: 00:20:41.723 { 00:20:41.723 "code": -32603, 00:20:41.723 "message": "Internal error" 00:20:41.723 } 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@181 -- # killprocess 3880152 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3880152 ']' 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3880152 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3880152 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3880152' 00:20:41.983 killing process with pid 3880152 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3880152 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3880152 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@182 -- # chmod 0600 /tmp/tmp.wq2lUTRnNY 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@185 -- # nvmfappstart -m 0x2 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3880584 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3880584 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3880584 ']' 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:41.983 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:41.983 19:16:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:41.983 [2024-11-26 19:16:54.592227] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:41.983 [2024-11-26 19:16:54.592280] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:42.245 [2024-11-26 19:16:54.687811] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:42.245 [2024-11-26 19:16:54.717911] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:42.245 [2024-11-26 19:16:54.717941] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:42.245 [2024-11-26 19:16:54.717947] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:42.245 [2024-11-26 19:16:54.717951] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:42.245 [2024-11-26 19:16:54.717955] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:42.245 [2024-11-26 19:16:54.718449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@186 -- # setup_nvmf_tgt /tmp/tmp.wq2lUTRnNY 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wq2lUTRnNY 00:20:42.814 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:20:43.074 [2024-11-26 19:16:55.571616] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:43.074 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:20:43.334 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:20:43.334 [2024-11-26 19:16:55.896421] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:20:43.334 [2024-11-26 19:16:55.896606] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:43.334 19:16:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:20:43.594 malloc0 00:20:43.594 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:20:43.854 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:20:43.854 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@188 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@189 -- # bdevperf_pid=3880975 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@191 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@192 -- # waitforlisten 3880975 /var/tmp/bdevperf.sock 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3880975 ']' 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:44.114 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:44.114 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:44.114 [2024-11-26 19:16:56.565992] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:44.114 [2024-11-26 19:16:56.566043] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3880975 ] 00:20:44.114 [2024-11-26 19:16:56.630253] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:44.114 [2024-11-26 19:16:56.659298] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:44.374 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:44.374 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:44.374 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@193 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:20:44.374 19:16:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@194 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:44.635 [2024-11-26 19:16:57.069473] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:44.635 TLSTESTn1 00:20:44.635 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py save_config 00:20:44.897 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # tgtconf='{ 00:20:44.897 "subsystems": [ 00:20:44.897 { 00:20:44.897 "subsystem": "keyring", 00:20:44.897 "config": [ 00:20:44.897 { 00:20:44.897 "method": "keyring_file_add_key", 00:20:44.897 "params": { 00:20:44.897 "name": "key0", 00:20:44.897 "path": "/tmp/tmp.wq2lUTRnNY" 00:20:44.897 } 00:20:44.897 } 00:20:44.897 ] 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "subsystem": "iobuf", 00:20:44.897 "config": [ 00:20:44.897 { 00:20:44.897 "method": "iobuf_set_options", 00:20:44.897 "params": { 00:20:44.897 "small_pool_count": 8192, 00:20:44.897 "large_pool_count": 1024, 00:20:44.897 "small_bufsize": 8192, 00:20:44.897 "large_bufsize": 135168, 00:20:44.897 "enable_numa": false 00:20:44.897 } 00:20:44.897 } 00:20:44.897 ] 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "subsystem": "sock", 00:20:44.897 "config": [ 00:20:44.897 { 00:20:44.897 "method": "sock_set_default_impl", 00:20:44.897 "params": { 00:20:44.897 "impl_name": "posix" 00:20:44.897 } 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "method": "sock_impl_set_options", 00:20:44.897 "params": { 00:20:44.897 "impl_name": "ssl", 00:20:44.897 "recv_buf_size": 4096, 00:20:44.897 "send_buf_size": 4096, 00:20:44.897 "enable_recv_pipe": true, 00:20:44.897 "enable_quickack": false, 00:20:44.897 "enable_placement_id": 0, 00:20:44.897 "enable_zerocopy_send_server": true, 00:20:44.897 "enable_zerocopy_send_client": false, 00:20:44.897 "zerocopy_threshold": 0, 00:20:44.897 "tls_version": 0, 00:20:44.897 "enable_ktls": false 00:20:44.897 } 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "method": "sock_impl_set_options", 00:20:44.897 "params": { 00:20:44.897 "impl_name": "posix", 00:20:44.897 "recv_buf_size": 2097152, 00:20:44.897 "send_buf_size": 2097152, 00:20:44.897 "enable_recv_pipe": true, 00:20:44.897 "enable_quickack": false, 00:20:44.897 "enable_placement_id": 0, 00:20:44.897 "enable_zerocopy_send_server": true, 00:20:44.897 "enable_zerocopy_send_client": false, 00:20:44.897 "zerocopy_threshold": 0, 00:20:44.897 "tls_version": 0, 00:20:44.897 "enable_ktls": false 00:20:44.897 } 00:20:44.897 } 00:20:44.897 ] 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "subsystem": "vmd", 00:20:44.897 "config": [] 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "subsystem": "accel", 00:20:44.897 "config": [ 00:20:44.897 { 00:20:44.897 "method": "accel_set_options", 00:20:44.897 "params": { 00:20:44.897 "small_cache_size": 128, 00:20:44.897 "large_cache_size": 16, 00:20:44.897 "task_count": 2048, 00:20:44.897 "sequence_count": 2048, 00:20:44.897 "buf_count": 2048 00:20:44.897 } 00:20:44.897 } 00:20:44.897 ] 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "subsystem": "bdev", 00:20:44.897 "config": [ 00:20:44.897 { 00:20:44.897 "method": "bdev_set_options", 00:20:44.897 "params": { 00:20:44.897 "bdev_io_pool_size": 65535, 00:20:44.897 "bdev_io_cache_size": 256, 00:20:44.897 "bdev_auto_examine": true, 00:20:44.897 "iobuf_small_cache_size": 128, 00:20:44.897 "iobuf_large_cache_size": 16 00:20:44.897 } 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "method": "bdev_raid_set_options", 00:20:44.897 "params": { 00:20:44.897 "process_window_size_kb": 1024, 00:20:44.897 "process_max_bandwidth_mb_sec": 0 00:20:44.897 } 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "method": "bdev_iscsi_set_options", 00:20:44.897 "params": { 00:20:44.897 "timeout_sec": 30 00:20:44.897 } 00:20:44.897 }, 00:20:44.897 { 00:20:44.897 "method": "bdev_nvme_set_options", 00:20:44.897 "params": { 00:20:44.897 "action_on_timeout": "none", 00:20:44.897 "timeout_us": 0, 00:20:44.897 "timeout_admin_us": 0, 00:20:44.897 "keep_alive_timeout_ms": 10000, 00:20:44.897 "arbitration_burst": 0, 00:20:44.897 "low_priority_weight": 0, 00:20:44.897 "medium_priority_weight": 0, 00:20:44.897 "high_priority_weight": 0, 00:20:44.897 "nvme_adminq_poll_period_us": 10000, 00:20:44.897 "nvme_ioq_poll_period_us": 0, 00:20:44.897 "io_queue_requests": 0, 00:20:44.897 "delay_cmd_submit": true, 00:20:44.897 "transport_retry_count": 4, 00:20:44.897 "bdev_retry_count": 3, 00:20:44.897 "transport_ack_timeout": 0, 00:20:44.897 "ctrlr_loss_timeout_sec": 0, 00:20:44.897 "reconnect_delay_sec": 0, 00:20:44.897 "fast_io_fail_timeout_sec": 0, 00:20:44.897 "disable_auto_failback": false, 00:20:44.897 "generate_uuids": false, 00:20:44.897 "transport_tos": 0, 00:20:44.897 "nvme_error_stat": false, 00:20:44.897 "rdma_srq_size": 0, 00:20:44.897 "io_path_stat": false, 00:20:44.897 "allow_accel_sequence": false, 00:20:44.897 "rdma_max_cq_size": 0, 00:20:44.897 "rdma_cm_event_timeout_ms": 0, 00:20:44.897 "dhchap_digests": [ 00:20:44.897 "sha256", 00:20:44.897 "sha384", 00:20:44.897 "sha512" 00:20:44.897 ], 00:20:44.897 "dhchap_dhgroups": [ 00:20:44.897 "null", 00:20:44.897 "ffdhe2048", 00:20:44.897 "ffdhe3072", 00:20:44.897 "ffdhe4096", 00:20:44.898 "ffdhe6144", 00:20:44.898 "ffdhe8192" 00:20:44.898 ] 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "bdev_nvme_set_hotplug", 00:20:44.898 "params": { 00:20:44.898 "period_us": 100000, 00:20:44.898 "enable": false 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "bdev_malloc_create", 00:20:44.898 "params": { 00:20:44.898 "name": "malloc0", 00:20:44.898 "num_blocks": 8192, 00:20:44.898 "block_size": 4096, 00:20:44.898 "physical_block_size": 4096, 00:20:44.898 "uuid": "bd16e55a-20ae-412e-ab3c-97fa929325a8", 00:20:44.898 "optimal_io_boundary": 0, 00:20:44.898 "md_size": 0, 00:20:44.898 "dif_type": 0, 00:20:44.898 "dif_is_head_of_md": false, 00:20:44.898 "dif_pi_format": 0 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "bdev_wait_for_examine" 00:20:44.898 } 00:20:44.898 ] 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "subsystem": "nbd", 00:20:44.898 "config": [] 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "subsystem": "scheduler", 00:20:44.898 "config": [ 00:20:44.898 { 00:20:44.898 "method": "framework_set_scheduler", 00:20:44.898 "params": { 00:20:44.898 "name": "static" 00:20:44.898 } 00:20:44.898 } 00:20:44.898 ] 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "subsystem": "nvmf", 00:20:44.898 "config": [ 00:20:44.898 { 00:20:44.898 "method": "nvmf_set_config", 00:20:44.898 "params": { 00:20:44.898 "discovery_filter": "match_any", 00:20:44.898 "admin_cmd_passthru": { 00:20:44.898 "identify_ctrlr": false 00:20:44.898 }, 00:20:44.898 "dhchap_digests": [ 00:20:44.898 "sha256", 00:20:44.898 "sha384", 00:20:44.898 "sha512" 00:20:44.898 ], 00:20:44.898 "dhchap_dhgroups": [ 00:20:44.898 "null", 00:20:44.898 "ffdhe2048", 00:20:44.898 "ffdhe3072", 00:20:44.898 "ffdhe4096", 00:20:44.898 "ffdhe6144", 00:20:44.898 "ffdhe8192" 00:20:44.898 ] 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "nvmf_set_max_subsystems", 00:20:44.898 "params": { 00:20:44.898 "max_subsystems": 1024 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "nvmf_set_crdt", 00:20:44.898 "params": { 00:20:44.898 "crdt1": 0, 00:20:44.898 "crdt2": 0, 00:20:44.898 "crdt3": 0 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "nvmf_create_transport", 00:20:44.898 "params": { 00:20:44.898 "trtype": "TCP", 00:20:44.898 "max_queue_depth": 128, 00:20:44.898 "max_io_qpairs_per_ctrlr": 127, 00:20:44.898 "in_capsule_data_size": 4096, 00:20:44.898 "max_io_size": 131072, 00:20:44.898 "io_unit_size": 131072, 00:20:44.898 "max_aq_depth": 128, 00:20:44.898 "num_shared_buffers": 511, 00:20:44.898 "buf_cache_size": 4294967295, 00:20:44.898 "dif_insert_or_strip": false, 00:20:44.898 "zcopy": false, 00:20:44.898 "c2h_success": false, 00:20:44.898 "sock_priority": 0, 00:20:44.898 "abort_timeout_sec": 1, 00:20:44.898 "ack_timeout": 0, 00:20:44.898 "data_wr_pool_size": 0 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "nvmf_create_subsystem", 00:20:44.898 "params": { 00:20:44.898 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:44.898 "allow_any_host": false, 00:20:44.898 "serial_number": "SPDK00000000000001", 00:20:44.898 "model_number": "SPDK bdev Controller", 00:20:44.898 "max_namespaces": 10, 00:20:44.898 "min_cntlid": 1, 00:20:44.898 "max_cntlid": 65519, 00:20:44.898 "ana_reporting": false 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "nvmf_subsystem_add_host", 00:20:44.898 "params": { 00:20:44.898 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:44.898 "host": "nqn.2016-06.io.spdk:host1", 00:20:44.898 "psk": "key0" 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "nvmf_subsystem_add_ns", 00:20:44.898 "params": { 00:20:44.898 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:44.898 "namespace": { 00:20:44.898 "nsid": 1, 00:20:44.898 "bdev_name": "malloc0", 00:20:44.898 "nguid": "BD16E55A20AE412EAB3C97FA929325A8", 00:20:44.898 "uuid": "bd16e55a-20ae-412e-ab3c-97fa929325a8", 00:20:44.898 "no_auto_visible": false 00:20:44.898 } 00:20:44.898 } 00:20:44.898 }, 00:20:44.898 { 00:20:44.898 "method": "nvmf_subsystem_add_listener", 00:20:44.898 "params": { 00:20:44.898 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:44.898 "listen_address": { 00:20:44.898 "trtype": "TCP", 00:20:44.898 "adrfam": "IPv4", 00:20:44.898 "traddr": "10.0.0.2", 00:20:44.898 "trsvcid": "4420" 00:20:44.898 }, 00:20:44.898 "secure_channel": true 00:20:44.898 } 00:20:44.898 } 00:20:44.898 ] 00:20:44.898 } 00:20:44.898 ] 00:20:44.898 }' 00:20:44.898 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:20:45.159 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # bdevperfconf='{ 00:20:45.159 "subsystems": [ 00:20:45.159 { 00:20:45.159 "subsystem": "keyring", 00:20:45.159 "config": [ 00:20:45.159 { 00:20:45.159 "method": "keyring_file_add_key", 00:20:45.159 "params": { 00:20:45.159 "name": "key0", 00:20:45.159 "path": "/tmp/tmp.wq2lUTRnNY" 00:20:45.159 } 00:20:45.159 } 00:20:45.159 ] 00:20:45.159 }, 00:20:45.159 { 00:20:45.159 "subsystem": "iobuf", 00:20:45.159 "config": [ 00:20:45.159 { 00:20:45.159 "method": "iobuf_set_options", 00:20:45.159 "params": { 00:20:45.159 "small_pool_count": 8192, 00:20:45.159 "large_pool_count": 1024, 00:20:45.159 "small_bufsize": 8192, 00:20:45.159 "large_bufsize": 135168, 00:20:45.159 "enable_numa": false 00:20:45.159 } 00:20:45.159 } 00:20:45.159 ] 00:20:45.159 }, 00:20:45.159 { 00:20:45.159 "subsystem": "sock", 00:20:45.159 "config": [ 00:20:45.159 { 00:20:45.159 "method": "sock_set_default_impl", 00:20:45.159 "params": { 00:20:45.159 "impl_name": "posix" 00:20:45.159 } 00:20:45.159 }, 00:20:45.159 { 00:20:45.160 "method": "sock_impl_set_options", 00:20:45.160 "params": { 00:20:45.160 "impl_name": "ssl", 00:20:45.160 "recv_buf_size": 4096, 00:20:45.160 "send_buf_size": 4096, 00:20:45.160 "enable_recv_pipe": true, 00:20:45.160 "enable_quickack": false, 00:20:45.160 "enable_placement_id": 0, 00:20:45.160 "enable_zerocopy_send_server": true, 00:20:45.160 "enable_zerocopy_send_client": false, 00:20:45.160 "zerocopy_threshold": 0, 00:20:45.160 "tls_version": 0, 00:20:45.160 "enable_ktls": false 00:20:45.160 } 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "method": "sock_impl_set_options", 00:20:45.160 "params": { 00:20:45.160 "impl_name": "posix", 00:20:45.160 "recv_buf_size": 2097152, 00:20:45.160 "send_buf_size": 2097152, 00:20:45.160 "enable_recv_pipe": true, 00:20:45.160 "enable_quickack": false, 00:20:45.160 "enable_placement_id": 0, 00:20:45.160 "enable_zerocopy_send_server": true, 00:20:45.160 "enable_zerocopy_send_client": false, 00:20:45.160 "zerocopy_threshold": 0, 00:20:45.160 "tls_version": 0, 00:20:45.160 "enable_ktls": false 00:20:45.160 } 00:20:45.160 } 00:20:45.160 ] 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "subsystem": "vmd", 00:20:45.160 "config": [] 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "subsystem": "accel", 00:20:45.160 "config": [ 00:20:45.160 { 00:20:45.160 "method": "accel_set_options", 00:20:45.160 "params": { 00:20:45.160 "small_cache_size": 128, 00:20:45.160 "large_cache_size": 16, 00:20:45.160 "task_count": 2048, 00:20:45.160 "sequence_count": 2048, 00:20:45.160 "buf_count": 2048 00:20:45.160 } 00:20:45.160 } 00:20:45.160 ] 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "subsystem": "bdev", 00:20:45.160 "config": [ 00:20:45.160 { 00:20:45.160 "method": "bdev_set_options", 00:20:45.160 "params": { 00:20:45.160 "bdev_io_pool_size": 65535, 00:20:45.160 "bdev_io_cache_size": 256, 00:20:45.160 "bdev_auto_examine": true, 00:20:45.160 "iobuf_small_cache_size": 128, 00:20:45.160 "iobuf_large_cache_size": 16 00:20:45.160 } 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "method": "bdev_raid_set_options", 00:20:45.160 "params": { 00:20:45.160 "process_window_size_kb": 1024, 00:20:45.160 "process_max_bandwidth_mb_sec": 0 00:20:45.160 } 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "method": "bdev_iscsi_set_options", 00:20:45.160 "params": { 00:20:45.160 "timeout_sec": 30 00:20:45.160 } 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "method": "bdev_nvme_set_options", 00:20:45.160 "params": { 00:20:45.160 "action_on_timeout": "none", 00:20:45.160 "timeout_us": 0, 00:20:45.160 "timeout_admin_us": 0, 00:20:45.160 "keep_alive_timeout_ms": 10000, 00:20:45.160 "arbitration_burst": 0, 00:20:45.160 "low_priority_weight": 0, 00:20:45.160 "medium_priority_weight": 0, 00:20:45.160 "high_priority_weight": 0, 00:20:45.160 "nvme_adminq_poll_period_us": 10000, 00:20:45.160 "nvme_ioq_poll_period_us": 0, 00:20:45.160 "io_queue_requests": 512, 00:20:45.160 "delay_cmd_submit": true, 00:20:45.160 "transport_retry_count": 4, 00:20:45.160 "bdev_retry_count": 3, 00:20:45.160 "transport_ack_timeout": 0, 00:20:45.160 "ctrlr_loss_timeout_sec": 0, 00:20:45.160 "reconnect_delay_sec": 0, 00:20:45.160 "fast_io_fail_timeout_sec": 0, 00:20:45.160 "disable_auto_failback": false, 00:20:45.160 "generate_uuids": false, 00:20:45.160 "transport_tos": 0, 00:20:45.160 "nvme_error_stat": false, 00:20:45.160 "rdma_srq_size": 0, 00:20:45.160 "io_path_stat": false, 00:20:45.160 "allow_accel_sequence": false, 00:20:45.160 "rdma_max_cq_size": 0, 00:20:45.160 "rdma_cm_event_timeout_ms": 0, 00:20:45.160 "dhchap_digests": [ 00:20:45.160 "sha256", 00:20:45.160 "sha384", 00:20:45.160 "sha512" 00:20:45.160 ], 00:20:45.160 "dhchap_dhgroups": [ 00:20:45.160 "null", 00:20:45.160 "ffdhe2048", 00:20:45.160 "ffdhe3072", 00:20:45.160 "ffdhe4096", 00:20:45.160 "ffdhe6144", 00:20:45.160 "ffdhe8192" 00:20:45.160 ] 00:20:45.160 } 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "method": "bdev_nvme_attach_controller", 00:20:45.160 "params": { 00:20:45.160 "name": "TLSTEST", 00:20:45.160 "trtype": "TCP", 00:20:45.160 "adrfam": "IPv4", 00:20:45.160 "traddr": "10.0.0.2", 00:20:45.160 "trsvcid": "4420", 00:20:45.160 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:45.160 "prchk_reftag": false, 00:20:45.160 "prchk_guard": false, 00:20:45.160 "ctrlr_loss_timeout_sec": 0, 00:20:45.160 "reconnect_delay_sec": 0, 00:20:45.160 "fast_io_fail_timeout_sec": 0, 00:20:45.160 "psk": "key0", 00:20:45.160 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:45.160 "hdgst": false, 00:20:45.160 "ddgst": false, 00:20:45.160 "multipath": "multipath" 00:20:45.160 } 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "method": "bdev_nvme_set_hotplug", 00:20:45.160 "params": { 00:20:45.160 "period_us": 100000, 00:20:45.160 "enable": false 00:20:45.160 } 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "method": "bdev_wait_for_examine" 00:20:45.160 } 00:20:45.160 ] 00:20:45.160 }, 00:20:45.160 { 00:20:45.160 "subsystem": "nbd", 00:20:45.160 "config": [] 00:20:45.160 } 00:20:45.160 ] 00:20:45.160 }' 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@201 -- # killprocess 3880975 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3880975 ']' 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3880975 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3880975 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3880975' 00:20:45.160 killing process with pid 3880975 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3880975 00:20:45.160 Received shutdown signal, test time was about 10.000000 seconds 00:20:45.160 00:20:45.160 Latency(us) 00:20:45.160 [2024-11-26T18:16:57.785Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:45.160 [2024-11-26T18:16:57.785Z] =================================================================================================================== 00:20:45.160 [2024-11-26T18:16:57.785Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:20:45.160 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3880975 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@202 -- # killprocess 3880584 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3880584 ']' 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3880584 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3880584 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3880584' 00:20:45.422 killing process with pid 3880584 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3880584 00:20:45.422 19:16:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3880584 00:20:45.422 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # nvmfappstart -m 0x2 -c /dev/fd/62 00:20:45.422 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:45.422 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:45.422 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:45.422 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # echo '{ 00:20:45.422 "subsystems": [ 00:20:45.422 { 00:20:45.422 "subsystem": "keyring", 00:20:45.422 "config": [ 00:20:45.422 { 00:20:45.422 "method": "keyring_file_add_key", 00:20:45.422 "params": { 00:20:45.422 "name": "key0", 00:20:45.422 "path": "/tmp/tmp.wq2lUTRnNY" 00:20:45.422 } 00:20:45.422 } 00:20:45.422 ] 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "subsystem": "iobuf", 00:20:45.422 "config": [ 00:20:45.422 { 00:20:45.422 "method": "iobuf_set_options", 00:20:45.422 "params": { 00:20:45.422 "small_pool_count": 8192, 00:20:45.422 "large_pool_count": 1024, 00:20:45.422 "small_bufsize": 8192, 00:20:45.422 "large_bufsize": 135168, 00:20:45.422 "enable_numa": false 00:20:45.422 } 00:20:45.422 } 00:20:45.422 ] 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "subsystem": "sock", 00:20:45.422 "config": [ 00:20:45.422 { 00:20:45.422 "method": "sock_set_default_impl", 00:20:45.422 "params": { 00:20:45.422 "impl_name": "posix" 00:20:45.422 } 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "method": "sock_impl_set_options", 00:20:45.422 "params": { 00:20:45.422 "impl_name": "ssl", 00:20:45.422 "recv_buf_size": 4096, 00:20:45.422 "send_buf_size": 4096, 00:20:45.422 "enable_recv_pipe": true, 00:20:45.422 "enable_quickack": false, 00:20:45.422 "enable_placement_id": 0, 00:20:45.422 "enable_zerocopy_send_server": true, 00:20:45.422 "enable_zerocopy_send_client": false, 00:20:45.422 "zerocopy_threshold": 0, 00:20:45.422 "tls_version": 0, 00:20:45.422 "enable_ktls": false 00:20:45.422 } 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "method": "sock_impl_set_options", 00:20:45.422 "params": { 00:20:45.422 "impl_name": "posix", 00:20:45.422 "recv_buf_size": 2097152, 00:20:45.422 "send_buf_size": 2097152, 00:20:45.422 "enable_recv_pipe": true, 00:20:45.422 "enable_quickack": false, 00:20:45.422 "enable_placement_id": 0, 00:20:45.422 "enable_zerocopy_send_server": true, 00:20:45.422 "enable_zerocopy_send_client": false, 00:20:45.422 "zerocopy_threshold": 0, 00:20:45.422 "tls_version": 0, 00:20:45.422 "enable_ktls": false 00:20:45.422 } 00:20:45.422 } 00:20:45.422 ] 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "subsystem": "vmd", 00:20:45.422 "config": [] 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "subsystem": "accel", 00:20:45.422 "config": [ 00:20:45.422 { 00:20:45.422 "method": "accel_set_options", 00:20:45.422 "params": { 00:20:45.422 "small_cache_size": 128, 00:20:45.422 "large_cache_size": 16, 00:20:45.422 "task_count": 2048, 00:20:45.422 "sequence_count": 2048, 00:20:45.422 "buf_count": 2048 00:20:45.422 } 00:20:45.422 } 00:20:45.422 ] 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "subsystem": "bdev", 00:20:45.422 "config": [ 00:20:45.422 { 00:20:45.422 "method": "bdev_set_options", 00:20:45.422 "params": { 00:20:45.422 "bdev_io_pool_size": 65535, 00:20:45.422 "bdev_io_cache_size": 256, 00:20:45.422 "bdev_auto_examine": true, 00:20:45.422 "iobuf_small_cache_size": 128, 00:20:45.422 "iobuf_large_cache_size": 16 00:20:45.422 } 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "method": "bdev_raid_set_options", 00:20:45.422 "params": { 00:20:45.422 "process_window_size_kb": 1024, 00:20:45.422 "process_max_bandwidth_mb_sec": 0 00:20:45.422 } 00:20:45.422 }, 00:20:45.422 { 00:20:45.422 "method": "bdev_iscsi_set_options", 00:20:45.423 "params": { 00:20:45.423 "timeout_sec": 30 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "bdev_nvme_set_options", 00:20:45.423 "params": { 00:20:45.423 "action_on_timeout": "none", 00:20:45.423 "timeout_us": 0, 00:20:45.423 "timeout_admin_us": 0, 00:20:45.423 "keep_alive_timeout_ms": 10000, 00:20:45.423 "arbitration_burst": 0, 00:20:45.423 "low_priority_weight": 0, 00:20:45.423 "medium_priority_weight": 0, 00:20:45.423 "high_priority_weight": 0, 00:20:45.423 "nvme_adminq_poll_period_us": 10000, 00:20:45.423 "nvme_ioq_poll_period_us": 0, 00:20:45.423 "io_queue_requests": 0, 00:20:45.423 "delay_cmd_submit": true, 00:20:45.423 "transport_retry_count": 4, 00:20:45.423 "bdev_retry_count": 3, 00:20:45.423 "transport_ack_timeout": 0, 00:20:45.423 "ctrlr_loss_timeout_sec": 0, 00:20:45.423 "reconnect_delay_sec": 0, 00:20:45.423 "fast_io_fail_timeout_sec": 0, 00:20:45.423 "disable_auto_failback": false, 00:20:45.423 "generate_uuids": false, 00:20:45.423 "transport_tos": 0, 00:20:45.423 "nvme_error_stat": false, 00:20:45.423 "rdma_srq_size": 0, 00:20:45.423 "io_path_stat": false, 00:20:45.423 "allow_accel_sequence": false, 00:20:45.423 "rdma_max_cq_size": 0, 00:20:45.423 "rdma_cm_event_timeout_ms": 0, 00:20:45.423 "dhchap_digests": [ 00:20:45.423 "sha256", 00:20:45.423 "sha384", 00:20:45.423 "sha512" 00:20:45.423 ], 00:20:45.423 "dhchap_dhgroups": [ 00:20:45.423 "null", 00:20:45.423 "ffdhe2048", 00:20:45.423 "ffdhe3072", 00:20:45.423 "ffdhe4096", 00:20:45.423 "ffdhe6144", 00:20:45.423 "ffdhe8192" 00:20:45.423 ] 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "bdev_nvme_set_hotplug", 00:20:45.423 "params": { 00:20:45.423 "period_us": 100000, 00:20:45.423 "enable": false 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "bdev_malloc_create", 00:20:45.423 "params": { 00:20:45.423 "name": "malloc0", 00:20:45.423 "num_blocks": 8192, 00:20:45.423 "block_size": 4096, 00:20:45.423 "physical_block_size": 4096, 00:20:45.423 "uuid": "bd16e55a-20ae-412e-ab3c-97fa929325a8", 00:20:45.423 "optimal_io_boundary": 0, 00:20:45.423 "md_size": 0, 00:20:45.423 "dif_type": 0, 00:20:45.423 "dif_is_head_of_md": false, 00:20:45.423 "dif_pi_format": 0 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "bdev_wait_for_examine" 00:20:45.423 } 00:20:45.423 ] 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "subsystem": "nbd", 00:20:45.423 "config": [] 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "subsystem": "scheduler", 00:20:45.423 "config": [ 00:20:45.423 { 00:20:45.423 "method": "framework_set_scheduler", 00:20:45.423 "params": { 00:20:45.423 "name": "static" 00:20:45.423 } 00:20:45.423 } 00:20:45.423 ] 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "subsystem": "nvmf", 00:20:45.423 "config": [ 00:20:45.423 { 00:20:45.423 "method": "nvmf_set_config", 00:20:45.423 "params": { 00:20:45.423 "discovery_filter": "match_any", 00:20:45.423 "admin_cmd_passthru": { 00:20:45.423 "identify_ctrlr": false 00:20:45.423 }, 00:20:45.423 "dhchap_digests": [ 00:20:45.423 "sha256", 00:20:45.423 "sha384", 00:20:45.423 "sha512" 00:20:45.423 ], 00:20:45.423 "dhchap_dhgroups": [ 00:20:45.423 "null", 00:20:45.423 "ffdhe2048", 00:20:45.423 "ffdhe3072", 00:20:45.423 "ffdhe4096", 00:20:45.423 "ffdhe6144", 00:20:45.423 "ffdhe8192" 00:20:45.423 ] 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "nvmf_set_max_subsystems", 00:20:45.423 "params": { 00:20:45.423 "max_subsystems": 1024 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "nvmf_set_crdt", 00:20:45.423 "params": { 00:20:45.423 "crdt1": 0, 00:20:45.423 "crdt2": 0, 00:20:45.423 "crdt3": 0 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "nvmf_create_transport", 00:20:45.423 "params": { 00:20:45.423 "trtype": "TCP", 00:20:45.423 "max_queue_depth": 128, 00:20:45.423 "max_io_qpairs_per_ctrlr": 127, 00:20:45.423 "in_capsule_data_size": 4096, 00:20:45.423 "max_io_size": 131072, 00:20:45.423 "io_unit_size": 131072, 00:20:45.423 "max_aq_depth": 128, 00:20:45.423 "num_shared_buffers": 511, 00:20:45.423 "buf_cache_size": 4294967295, 00:20:45.423 "dif_insert_or_strip": false, 00:20:45.423 "zcopy": false, 00:20:45.423 "c2h_success": false, 00:20:45.423 "sock_priority": 0, 00:20:45.423 "abort_timeout_sec": 1, 00:20:45.423 "ack_timeout": 0, 00:20:45.423 "data_wr_pool_size": 0 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "nvmf_create_subsystem", 00:20:45.423 "params": { 00:20:45.423 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:45.423 "allow_any_host": false, 00:20:45.423 "serial_number": "SPDK00000000000001", 00:20:45.423 "model_number": "SPDK bdev Controller", 00:20:45.423 "max_namespaces": 10, 00:20:45.423 "min_cntlid": 1, 00:20:45.423 "max_cntlid": 65519, 00:20:45.423 "ana_reporting": false 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "nvmf_subsystem_add_host", 00:20:45.423 "params": { 00:20:45.423 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:45.423 "host": "nqn.2016-06.io.spdk:host1", 00:20:45.423 "psk": "key0" 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "nvmf_subsystem_add_ns", 00:20:45.423 "params": { 00:20:45.423 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:45.423 "namespace": { 00:20:45.423 "nsid": 1, 00:20:45.423 "bdev_name": "malloc0", 00:20:45.423 "nguid": "BD16E55A20AE412EAB3C97FA929325A8", 00:20:45.423 "uuid": "bd16e55a-20ae-412e-ab3c-97fa929325a8", 00:20:45.423 "no_auto_visible": false 00:20:45.423 } 00:20:45.423 } 00:20:45.423 }, 00:20:45.423 { 00:20:45.423 "method": "nvmf_subsystem_add_listener", 00:20:45.423 "params": { 00:20:45.423 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:45.423 "listen_address": { 00:20:45.423 "trtype": "TCP", 00:20:45.423 "adrfam": "IPv4", 00:20:45.423 "traddr": "10.0.0.2", 00:20:45.423 "trsvcid": "4420" 00:20:45.423 }, 00:20:45.423 "secure_channel": true 00:20:45.423 } 00:20:45.423 } 00:20:45.423 ] 00:20:45.423 } 00:20:45.423 ] 00:20:45.423 }' 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3881344 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3881344 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 -c /dev/fd/62 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3881344 ']' 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:45.423 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:45.423 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:45.683 [2024-11-26 19:16:58.079801] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:45.683 [2024-11-26 19:16:58.079859] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:45.683 [2024-11-26 19:16:58.177589] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:45.684 [2024-11-26 19:16:58.206892] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:45.684 [2024-11-26 19:16:58.206919] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:45.684 [2024-11-26 19:16:58.206925] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:45.684 [2024-11-26 19:16:58.206930] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:45.684 [2024-11-26 19:16:58.206934] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:45.684 [2024-11-26 19:16:58.207424] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:45.944 [2024-11-26 19:16:58.401401] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:45.944 [2024-11-26 19:16:58.433432] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:20:45.944 [2024-11-26 19:16:58.433623] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@209 -- # bdevperf_pid=3881494 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@210 -- # waitforlisten 3881494 /var/tmp/bdevperf.sock 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3881494 ']' 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:46.516 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 -c /dev/fd/63 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:46.516 19:16:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # echo '{ 00:20:46.516 "subsystems": [ 00:20:46.516 { 00:20:46.516 "subsystem": "keyring", 00:20:46.516 "config": [ 00:20:46.516 { 00:20:46.516 "method": "keyring_file_add_key", 00:20:46.516 "params": { 00:20:46.516 "name": "key0", 00:20:46.516 "path": "/tmp/tmp.wq2lUTRnNY" 00:20:46.516 } 00:20:46.516 } 00:20:46.516 ] 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "subsystem": "iobuf", 00:20:46.516 "config": [ 00:20:46.516 { 00:20:46.516 "method": "iobuf_set_options", 00:20:46.516 "params": { 00:20:46.516 "small_pool_count": 8192, 00:20:46.516 "large_pool_count": 1024, 00:20:46.516 "small_bufsize": 8192, 00:20:46.516 "large_bufsize": 135168, 00:20:46.516 "enable_numa": false 00:20:46.516 } 00:20:46.516 } 00:20:46.516 ] 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "subsystem": "sock", 00:20:46.516 "config": [ 00:20:46.516 { 00:20:46.516 "method": "sock_set_default_impl", 00:20:46.516 "params": { 00:20:46.516 "impl_name": "posix" 00:20:46.516 } 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "method": "sock_impl_set_options", 00:20:46.516 "params": { 00:20:46.516 "impl_name": "ssl", 00:20:46.516 "recv_buf_size": 4096, 00:20:46.516 "send_buf_size": 4096, 00:20:46.516 "enable_recv_pipe": true, 00:20:46.516 "enable_quickack": false, 00:20:46.516 "enable_placement_id": 0, 00:20:46.516 "enable_zerocopy_send_server": true, 00:20:46.516 "enable_zerocopy_send_client": false, 00:20:46.516 "zerocopy_threshold": 0, 00:20:46.516 "tls_version": 0, 00:20:46.516 "enable_ktls": false 00:20:46.516 } 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "method": "sock_impl_set_options", 00:20:46.516 "params": { 00:20:46.516 "impl_name": "posix", 00:20:46.516 "recv_buf_size": 2097152, 00:20:46.516 "send_buf_size": 2097152, 00:20:46.516 "enable_recv_pipe": true, 00:20:46.516 "enable_quickack": false, 00:20:46.516 "enable_placement_id": 0, 00:20:46.516 "enable_zerocopy_send_server": true, 00:20:46.516 "enable_zerocopy_send_client": false, 00:20:46.516 "zerocopy_threshold": 0, 00:20:46.516 "tls_version": 0, 00:20:46.516 "enable_ktls": false 00:20:46.516 } 00:20:46.516 } 00:20:46.516 ] 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "subsystem": "vmd", 00:20:46.516 "config": [] 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "subsystem": "accel", 00:20:46.516 "config": [ 00:20:46.516 { 00:20:46.516 "method": "accel_set_options", 00:20:46.516 "params": { 00:20:46.516 "small_cache_size": 128, 00:20:46.516 "large_cache_size": 16, 00:20:46.516 "task_count": 2048, 00:20:46.516 "sequence_count": 2048, 00:20:46.516 "buf_count": 2048 00:20:46.516 } 00:20:46.516 } 00:20:46.516 ] 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "subsystem": "bdev", 00:20:46.516 "config": [ 00:20:46.516 { 00:20:46.516 "method": "bdev_set_options", 00:20:46.516 "params": { 00:20:46.516 "bdev_io_pool_size": 65535, 00:20:46.516 "bdev_io_cache_size": 256, 00:20:46.516 "bdev_auto_examine": true, 00:20:46.516 "iobuf_small_cache_size": 128, 00:20:46.516 "iobuf_large_cache_size": 16 00:20:46.516 } 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "method": "bdev_raid_set_options", 00:20:46.516 "params": { 00:20:46.516 "process_window_size_kb": 1024, 00:20:46.516 "process_max_bandwidth_mb_sec": 0 00:20:46.516 } 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "method": "bdev_iscsi_set_options", 00:20:46.516 "params": { 00:20:46.516 "timeout_sec": 30 00:20:46.516 } 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "method": "bdev_nvme_set_options", 00:20:46.516 "params": { 00:20:46.516 "action_on_timeout": "none", 00:20:46.516 "timeout_us": 0, 00:20:46.516 "timeout_admin_us": 0, 00:20:46.516 "keep_alive_timeout_ms": 10000, 00:20:46.516 "arbitration_burst": 0, 00:20:46.516 "low_priority_weight": 0, 00:20:46.516 "medium_priority_weight": 0, 00:20:46.516 "high_priority_weight": 0, 00:20:46.516 "nvme_adminq_poll_period_us": 10000, 00:20:46.516 "nvme_ioq_poll_period_us": 0, 00:20:46.516 "io_queue_requests": 512, 00:20:46.516 "delay_cmd_submit": true, 00:20:46.516 "transport_retry_count": 4, 00:20:46.516 "bdev_retry_count": 3, 00:20:46.516 "transport_ack_timeout": 0, 00:20:46.516 "ctrlr_loss_timeout_sec": 0, 00:20:46.516 "reconnect_delay_sec": 0, 00:20:46.516 "fast_io_fail_timeout_sec": 0, 00:20:46.516 "disable_auto_failback": false, 00:20:46.516 "generate_uuids": false, 00:20:46.516 "transport_tos": 0, 00:20:46.516 "nvme_error_stat": false, 00:20:46.516 "rdma_srq_size": 0, 00:20:46.516 "io_path_stat": false, 00:20:46.516 "allow_accel_sequence": false, 00:20:46.516 "rdma_max_cq_size": 0, 00:20:46.516 "rdma_cm_event_timeout_ms": 0, 00:20:46.516 "dhchap_digests": [ 00:20:46.516 "sha256", 00:20:46.516 "sha384", 00:20:46.516 "sha512" 00:20:46.516 ], 00:20:46.516 "dhchap_dhgroups": [ 00:20:46.516 "null", 00:20:46.516 "ffdhe2048", 00:20:46.516 "ffdhe3072", 00:20:46.516 "ffdhe4096", 00:20:46.516 "ffdhe6144", 00:20:46.516 "ffdhe8192" 00:20:46.516 ] 00:20:46.516 } 00:20:46.516 }, 00:20:46.516 { 00:20:46.516 "method": "bdev_nvme_attach_controller", 00:20:46.516 "params": { 00:20:46.516 "name": "TLSTEST", 00:20:46.516 "trtype": "TCP", 00:20:46.516 "adrfam": "IPv4", 00:20:46.516 "traddr": "10.0.0.2", 00:20:46.517 "trsvcid": "4420", 00:20:46.517 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:46.517 "prchk_reftag": false, 00:20:46.517 "prchk_guard": false, 00:20:46.517 "ctrlr_loss_timeout_sec": 0, 00:20:46.517 "reconnect_delay_sec": 0, 00:20:46.517 "fast_io_fail_timeout_sec": 0, 00:20:46.517 "psk": "key0", 00:20:46.517 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:46.517 "hdgst": false, 00:20:46.517 "ddgst": false, 00:20:46.517 "multipath": "multipath" 00:20:46.517 } 00:20:46.517 }, 00:20:46.517 { 00:20:46.517 "method": "bdev_nvme_set_hotplug", 00:20:46.517 "params": { 00:20:46.517 "period_us": 100000, 00:20:46.517 "enable": false 00:20:46.517 } 00:20:46.517 }, 00:20:46.517 { 00:20:46.517 "method": "bdev_wait_for_examine" 00:20:46.517 } 00:20:46.517 ] 00:20:46.517 }, 00:20:46.517 { 00:20:46.517 "subsystem": "nbd", 00:20:46.517 "config": [] 00:20:46.517 } 00:20:46.517 ] 00:20:46.517 }' 00:20:46.517 [2024-11-26 19:16:58.956602] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:46.517 [2024-11-26 19:16:58.956653] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3881494 ] 00:20:46.517 [2024-11-26 19:16:59.019987] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:46.517 [2024-11-26 19:16:59.049167] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:46.777 [2024-11-26 19:16:59.184207] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:47.346 19:16:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:47.346 19:16:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:47.346 19:16:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@213 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:20:47.346 Running I/O for 10 seconds... 00:20:49.233 5965.00 IOPS, 23.30 MiB/s [2024-11-26T18:17:03.242Z] 5734.50 IOPS, 22.40 MiB/s [2024-11-26T18:17:04.182Z] 5773.00 IOPS, 22.55 MiB/s [2024-11-26T18:17:05.121Z] 5764.00 IOPS, 22.52 MiB/s [2024-11-26T18:17:06.064Z] 5774.00 IOPS, 22.55 MiB/s [2024-11-26T18:17:07.004Z] 5812.67 IOPS, 22.71 MiB/s [2024-11-26T18:17:07.946Z] 5841.14 IOPS, 22.82 MiB/s [2024-11-26T18:17:08.886Z] 5646.75 IOPS, 22.06 MiB/s [2024-11-26T18:17:09.873Z] 5586.44 IOPS, 21.82 MiB/s [2024-11-26T18:17:10.168Z] 5519.80 IOPS, 21.56 MiB/s 00:20:57.543 Latency(us) 00:20:57.543 [2024-11-26T18:17:10.168Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:57.543 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:20:57.543 Verification LBA range: start 0x0 length 0x2000 00:20:57.543 TLSTESTn1 : 10.04 5512.92 21.53 0.00 0.00 23173.36 4696.75 66409.81 00:20:57.543 [2024-11-26T18:17:10.168Z] =================================================================================================================== 00:20:57.543 [2024-11-26T18:17:10.168Z] Total : 5512.92 21.53 0.00 0.00 23173.36 4696.75 66409.81 00:20:57.543 { 00:20:57.543 "results": [ 00:20:57.543 { 00:20:57.543 "job": "TLSTESTn1", 00:20:57.543 "core_mask": "0x4", 00:20:57.543 "workload": "verify", 00:20:57.543 "status": "finished", 00:20:57.543 "verify_range": { 00:20:57.543 "start": 0, 00:20:57.543 "length": 8192 00:20:57.543 }, 00:20:57.543 "queue_depth": 128, 00:20:57.543 "io_size": 4096, 00:20:57.543 "runtime": 10.035697, 00:20:57.543 "iops": 5512.920527592652, 00:20:57.543 "mibps": 21.5348458109088, 00:20:57.543 "io_failed": 0, 00:20:57.543 "io_timeout": 0, 00:20:57.543 "avg_latency_us": 23173.36221426936, 00:20:57.543 "min_latency_us": 4696.746666666667, 00:20:57.543 "max_latency_us": 66409.81333333334 00:20:57.543 } 00:20:57.543 ], 00:20:57.543 "core_count": 1 00:20:57.543 } 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@215 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@216 -- # killprocess 3881494 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3881494 ']' 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3881494 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3881494 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3881494' 00:20:57.543 killing process with pid 3881494 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3881494 00:20:57.543 Received shutdown signal, test time was about 10.000000 seconds 00:20:57.543 00:20:57.543 Latency(us) 00:20:57.543 [2024-11-26T18:17:10.168Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:57.543 [2024-11-26T18:17:10.168Z] =================================================================================================================== 00:20:57.543 [2024-11-26T18:17:10.168Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:20:57.543 19:17:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3881494 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@217 -- # killprocess 3881344 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3881344 ']' 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3881344 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3881344 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3881344' 00:20:57.543 killing process with pid 3881344 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3881344 00:20:57.543 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3881344 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@220 -- # nvmfappstart 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3883733 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3883733 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3883733 ']' 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:57.805 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:57.805 19:17:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:57.805 [2024-11-26 19:17:10.303437] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:20:57.806 [2024-11-26 19:17:10.303491] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:57.806 [2024-11-26 19:17:10.390106] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:57.806 [2024-11-26 19:17:10.424814] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:57.806 [2024-11-26 19:17:10.424849] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:57.806 [2024-11-26 19:17:10.424858] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:57.806 [2024-11-26 19:17:10.424870] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:57.806 [2024-11-26 19:17:10.424876] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:57.806 [2024-11-26 19:17:10.425418] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@221 -- # setup_nvmf_tgt /tmp/tmp.wq2lUTRnNY 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wq2lUTRnNY 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:20:58.748 [2024-11-26 19:17:11.294932] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:58.748 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:20:59.008 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:20:59.268 [2024-11-26 19:17:11.655837] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:20:59.268 [2024-11-26 19:17:11.656052] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:59.268 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:20:59.268 malloc0 00:20:59.268 19:17:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:20:59.529 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@224 -- # bdevperf_pid=3884108 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@226 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@222 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@227 -- # waitforlisten 3884108 /var/tmp/bdevperf.sock 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3884108 ']' 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:59.789 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:59.789 19:17:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:00.049 [2024-11-26 19:17:12.446999] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:00.049 [2024-11-26 19:17:12.447054] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3884108 ] 00:21:00.049 [2024-11-26 19:17:12.535704] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:00.049 [2024-11-26 19:17:12.565747] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:00.619 19:17:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:00.619 19:17:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:21:00.619 19:17:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@229 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:21:00.880 19:17:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@230 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:21:01.141 [2024-11-26 19:17:13.522705] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:21:01.141 nvme0n1 00:21:01.141 19:17:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@234 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:21:01.141 Running I/O for 1 seconds... 00:21:02.342 5489.00 IOPS, 21.44 MiB/s 00:21:02.342 Latency(us) 00:21:02.342 [2024-11-26T18:17:14.967Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:02.342 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:21:02.342 Verification LBA range: start 0x0 length 0x2000 00:21:02.342 nvme0n1 : 1.02 5504.49 21.50 0.00 0.00 23027.36 4560.21 27088.21 00:21:02.342 [2024-11-26T18:17:14.967Z] =================================================================================================================== 00:21:02.342 [2024-11-26T18:17:14.967Z] Total : 5504.49 21.50 0.00 0.00 23027.36 4560.21 27088.21 00:21:02.342 { 00:21:02.342 "results": [ 00:21:02.342 { 00:21:02.342 "job": "nvme0n1", 00:21:02.342 "core_mask": "0x2", 00:21:02.342 "workload": "verify", 00:21:02.342 "status": "finished", 00:21:02.342 "verify_range": { 00:21:02.342 "start": 0, 00:21:02.342 "length": 8192 00:21:02.342 }, 00:21:02.342 "queue_depth": 128, 00:21:02.342 "io_size": 4096, 00:21:02.342 "runtime": 1.020621, 00:21:02.342 "iops": 5504.491873085112, 00:21:02.342 "mibps": 21.50192137923872, 00:21:02.342 "io_failed": 0, 00:21:02.342 "io_timeout": 0, 00:21:02.342 "avg_latency_us": 23027.36495550018, 00:21:02.342 "min_latency_us": 4560.213333333333, 00:21:02.342 "max_latency_us": 27088.213333333333 00:21:02.342 } 00:21:02.342 ], 00:21:02.342 "core_count": 1 00:21:02.342 } 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@236 -- # killprocess 3884108 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3884108 ']' 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3884108 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3884108 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3884108' 00:21:02.342 killing process with pid 3884108 00:21:02.342 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3884108 00:21:02.342 Received shutdown signal, test time was about 1.000000 seconds 00:21:02.342 00:21:02.343 Latency(us) 00:21:02.343 [2024-11-26T18:17:14.968Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:02.343 [2024-11-26T18:17:14.968Z] =================================================================================================================== 00:21:02.343 [2024-11-26T18:17:14.968Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3884108 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@237 -- # killprocess 3883733 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3883733 ']' 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3883733 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3883733 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3883733' 00:21:02.343 killing process with pid 3883733 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3883733 00:21:02.343 19:17:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3883733 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@242 -- # nvmfappstart 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3884784 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3884784 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3884784 ']' 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:02.604 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:02.604 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:02.604 [2024-11-26 19:17:15.148521] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:02.604 [2024-11-26 19:17:15.148574] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:02.872 [2024-11-26 19:17:15.235493] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:02.872 [2024-11-26 19:17:15.270083] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:02.872 [2024-11-26 19:17:15.270121] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:02.872 [2024-11-26 19:17:15.270129] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:02.872 [2024-11-26 19:17:15.270136] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:02.872 [2024-11-26 19:17:15.270141] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:02.872 [2024-11-26 19:17:15.270728] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@243 -- # rpc_cmd 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:03.448 19:17:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:03.448 [2024-11-26 19:17:15.996019] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:03.448 malloc0 00:21:03.448 [2024-11-26 19:17:16.022604] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:21:03.448 [2024-11-26 19:17:16.022822] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@256 -- # bdevperf_pid=3884850 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@258 -- # waitforlisten 3884850 /var/tmp/bdevperf.sock 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@254 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3884850 ']' 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:03.448 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:03.448 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:03.709 [2024-11-26 19:17:16.102059] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:03.709 [2024-11-26 19:17:16.102109] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3884850 ] 00:21:03.709 [2024-11-26 19:17:16.192024] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:03.709 [2024-11-26 19:17:16.222169] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:04.280 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:04.280 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:21:04.280 19:17:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@259 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wq2lUTRnNY 00:21:04.540 19:17:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@260 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:21:04.801 [2024-11-26 19:17:17.222948] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:21:04.801 nvme0n1 00:21:04.801 19:17:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@264 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:21:04.801 Running I/O for 1 seconds... 00:21:06.186 4015.00 IOPS, 15.68 MiB/s 00:21:06.186 Latency(us) 00:21:06.186 [2024-11-26T18:17:18.811Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:06.186 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:21:06.186 Verification LBA range: start 0x0 length 0x2000 00:21:06.186 nvme0n1 : 1.02 4060.12 15.86 0.00 0.00 31249.65 6335.15 24029.87 00:21:06.186 [2024-11-26T18:17:18.811Z] =================================================================================================================== 00:21:06.186 [2024-11-26T18:17:18.812Z] Total : 4060.12 15.86 0.00 0.00 31249.65 6335.15 24029.87 00:21:06.187 { 00:21:06.187 "results": [ 00:21:06.187 { 00:21:06.187 "job": "nvme0n1", 00:21:06.187 "core_mask": "0x2", 00:21:06.187 "workload": "verify", 00:21:06.187 "status": "finished", 00:21:06.187 "verify_range": { 00:21:06.187 "start": 0, 00:21:06.187 "length": 8192 00:21:06.187 }, 00:21:06.187 "queue_depth": 128, 00:21:06.187 "io_size": 4096, 00:21:06.187 "runtime": 1.02066, 00:21:06.187 "iops": 4060.1179628867594, 00:21:06.187 "mibps": 15.859835792526404, 00:21:06.187 "io_failed": 0, 00:21:06.187 "io_timeout": 0, 00:21:06.187 "avg_latency_us": 31249.653127413127, 00:21:06.187 "min_latency_us": 6335.1466666666665, 00:21:06.187 "max_latency_us": 24029.866666666665 00:21:06.187 } 00:21:06.187 ], 00:21:06.187 "core_count": 1 00:21:06.187 } 00:21:06.187 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # rpc_cmd save_config 00:21:06.187 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:06.187 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:06.187 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:06.187 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # tgtcfg='{ 00:21:06.187 "subsystems": [ 00:21:06.187 { 00:21:06.187 "subsystem": "keyring", 00:21:06.187 "config": [ 00:21:06.187 { 00:21:06.187 "method": "keyring_file_add_key", 00:21:06.187 "params": { 00:21:06.187 "name": "key0", 00:21:06.187 "path": "/tmp/tmp.wq2lUTRnNY" 00:21:06.187 } 00:21:06.187 } 00:21:06.187 ] 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "subsystem": "iobuf", 00:21:06.187 "config": [ 00:21:06.187 { 00:21:06.187 "method": "iobuf_set_options", 00:21:06.187 "params": { 00:21:06.187 "small_pool_count": 8192, 00:21:06.187 "large_pool_count": 1024, 00:21:06.187 "small_bufsize": 8192, 00:21:06.187 "large_bufsize": 135168, 00:21:06.187 "enable_numa": false 00:21:06.187 } 00:21:06.187 } 00:21:06.187 ] 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "subsystem": "sock", 00:21:06.187 "config": [ 00:21:06.187 { 00:21:06.187 "method": "sock_set_default_impl", 00:21:06.187 "params": { 00:21:06.187 "impl_name": "posix" 00:21:06.187 } 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "method": "sock_impl_set_options", 00:21:06.187 "params": { 00:21:06.187 "impl_name": "ssl", 00:21:06.187 "recv_buf_size": 4096, 00:21:06.187 "send_buf_size": 4096, 00:21:06.187 "enable_recv_pipe": true, 00:21:06.187 "enable_quickack": false, 00:21:06.187 "enable_placement_id": 0, 00:21:06.187 "enable_zerocopy_send_server": true, 00:21:06.187 "enable_zerocopy_send_client": false, 00:21:06.187 "zerocopy_threshold": 0, 00:21:06.187 "tls_version": 0, 00:21:06.187 "enable_ktls": false 00:21:06.187 } 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "method": "sock_impl_set_options", 00:21:06.187 "params": { 00:21:06.187 "impl_name": "posix", 00:21:06.187 "recv_buf_size": 2097152, 00:21:06.187 "send_buf_size": 2097152, 00:21:06.187 "enable_recv_pipe": true, 00:21:06.187 "enable_quickack": false, 00:21:06.187 "enable_placement_id": 0, 00:21:06.187 "enable_zerocopy_send_server": true, 00:21:06.187 "enable_zerocopy_send_client": false, 00:21:06.187 "zerocopy_threshold": 0, 00:21:06.187 "tls_version": 0, 00:21:06.187 "enable_ktls": false 00:21:06.187 } 00:21:06.187 } 00:21:06.187 ] 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "subsystem": "vmd", 00:21:06.187 "config": [] 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "subsystem": "accel", 00:21:06.187 "config": [ 00:21:06.187 { 00:21:06.187 "method": "accel_set_options", 00:21:06.187 "params": { 00:21:06.187 "small_cache_size": 128, 00:21:06.187 "large_cache_size": 16, 00:21:06.187 "task_count": 2048, 00:21:06.187 "sequence_count": 2048, 00:21:06.187 "buf_count": 2048 00:21:06.187 } 00:21:06.187 } 00:21:06.187 ] 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "subsystem": "bdev", 00:21:06.187 "config": [ 00:21:06.187 { 00:21:06.187 "method": "bdev_set_options", 00:21:06.187 "params": { 00:21:06.187 "bdev_io_pool_size": 65535, 00:21:06.187 "bdev_io_cache_size": 256, 00:21:06.187 "bdev_auto_examine": true, 00:21:06.187 "iobuf_small_cache_size": 128, 00:21:06.187 "iobuf_large_cache_size": 16 00:21:06.187 } 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "method": "bdev_raid_set_options", 00:21:06.187 "params": { 00:21:06.187 "process_window_size_kb": 1024, 00:21:06.187 "process_max_bandwidth_mb_sec": 0 00:21:06.187 } 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "method": "bdev_iscsi_set_options", 00:21:06.187 "params": { 00:21:06.187 "timeout_sec": 30 00:21:06.187 } 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "method": "bdev_nvme_set_options", 00:21:06.187 "params": { 00:21:06.187 "action_on_timeout": "none", 00:21:06.187 "timeout_us": 0, 00:21:06.187 "timeout_admin_us": 0, 00:21:06.187 "keep_alive_timeout_ms": 10000, 00:21:06.187 "arbitration_burst": 0, 00:21:06.187 "low_priority_weight": 0, 00:21:06.187 "medium_priority_weight": 0, 00:21:06.187 "high_priority_weight": 0, 00:21:06.187 "nvme_adminq_poll_period_us": 10000, 00:21:06.187 "nvme_ioq_poll_period_us": 0, 00:21:06.187 "io_queue_requests": 0, 00:21:06.187 "delay_cmd_submit": true, 00:21:06.187 "transport_retry_count": 4, 00:21:06.187 "bdev_retry_count": 3, 00:21:06.187 "transport_ack_timeout": 0, 00:21:06.187 "ctrlr_loss_timeout_sec": 0, 00:21:06.187 "reconnect_delay_sec": 0, 00:21:06.187 "fast_io_fail_timeout_sec": 0, 00:21:06.187 "disable_auto_failback": false, 00:21:06.187 "generate_uuids": false, 00:21:06.187 "transport_tos": 0, 00:21:06.187 "nvme_error_stat": false, 00:21:06.187 "rdma_srq_size": 0, 00:21:06.187 "io_path_stat": false, 00:21:06.187 "allow_accel_sequence": false, 00:21:06.187 "rdma_max_cq_size": 0, 00:21:06.187 "rdma_cm_event_timeout_ms": 0, 00:21:06.187 "dhchap_digests": [ 00:21:06.187 "sha256", 00:21:06.187 "sha384", 00:21:06.187 "sha512" 00:21:06.187 ], 00:21:06.187 "dhchap_dhgroups": [ 00:21:06.187 "null", 00:21:06.187 "ffdhe2048", 00:21:06.187 "ffdhe3072", 00:21:06.187 "ffdhe4096", 00:21:06.187 "ffdhe6144", 00:21:06.187 "ffdhe8192" 00:21:06.187 ] 00:21:06.187 } 00:21:06.187 }, 00:21:06.187 { 00:21:06.187 "method": "bdev_nvme_set_hotplug", 00:21:06.187 "params": { 00:21:06.187 "period_us": 100000, 00:21:06.188 "enable": false 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "bdev_malloc_create", 00:21:06.188 "params": { 00:21:06.188 "name": "malloc0", 00:21:06.188 "num_blocks": 8192, 00:21:06.188 "block_size": 4096, 00:21:06.188 "physical_block_size": 4096, 00:21:06.188 "uuid": "43318b22-96e3-48c3-a987-b6e33f7ebd6a", 00:21:06.188 "optimal_io_boundary": 0, 00:21:06.188 "md_size": 0, 00:21:06.188 "dif_type": 0, 00:21:06.188 "dif_is_head_of_md": false, 00:21:06.188 "dif_pi_format": 0 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "bdev_wait_for_examine" 00:21:06.188 } 00:21:06.188 ] 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "subsystem": "nbd", 00:21:06.188 "config": [] 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "subsystem": "scheduler", 00:21:06.188 "config": [ 00:21:06.188 { 00:21:06.188 "method": "framework_set_scheduler", 00:21:06.188 "params": { 00:21:06.188 "name": "static" 00:21:06.188 } 00:21:06.188 } 00:21:06.188 ] 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "subsystem": "nvmf", 00:21:06.188 "config": [ 00:21:06.188 { 00:21:06.188 "method": "nvmf_set_config", 00:21:06.188 "params": { 00:21:06.188 "discovery_filter": "match_any", 00:21:06.188 "admin_cmd_passthru": { 00:21:06.188 "identify_ctrlr": false 00:21:06.188 }, 00:21:06.188 "dhchap_digests": [ 00:21:06.188 "sha256", 00:21:06.188 "sha384", 00:21:06.188 "sha512" 00:21:06.188 ], 00:21:06.188 "dhchap_dhgroups": [ 00:21:06.188 "null", 00:21:06.188 "ffdhe2048", 00:21:06.188 "ffdhe3072", 00:21:06.188 "ffdhe4096", 00:21:06.188 "ffdhe6144", 00:21:06.188 "ffdhe8192" 00:21:06.188 ] 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "nvmf_set_max_subsystems", 00:21:06.188 "params": { 00:21:06.188 "max_subsystems": 1024 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "nvmf_set_crdt", 00:21:06.188 "params": { 00:21:06.188 "crdt1": 0, 00:21:06.188 "crdt2": 0, 00:21:06.188 "crdt3": 0 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "nvmf_create_transport", 00:21:06.188 "params": { 00:21:06.188 "trtype": "TCP", 00:21:06.188 "max_queue_depth": 128, 00:21:06.188 "max_io_qpairs_per_ctrlr": 127, 00:21:06.188 "in_capsule_data_size": 4096, 00:21:06.188 "max_io_size": 131072, 00:21:06.188 "io_unit_size": 131072, 00:21:06.188 "max_aq_depth": 128, 00:21:06.188 "num_shared_buffers": 511, 00:21:06.188 "buf_cache_size": 4294967295, 00:21:06.188 "dif_insert_or_strip": false, 00:21:06.188 "zcopy": false, 00:21:06.188 "c2h_success": false, 00:21:06.188 "sock_priority": 0, 00:21:06.188 "abort_timeout_sec": 1, 00:21:06.188 "ack_timeout": 0, 00:21:06.188 "data_wr_pool_size": 0 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "nvmf_create_subsystem", 00:21:06.188 "params": { 00:21:06.188 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.188 "allow_any_host": false, 00:21:06.188 "serial_number": "00000000000000000000", 00:21:06.188 "model_number": "SPDK bdev Controller", 00:21:06.188 "max_namespaces": 32, 00:21:06.188 "min_cntlid": 1, 00:21:06.188 "max_cntlid": 65519, 00:21:06.188 "ana_reporting": false 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "nvmf_subsystem_add_host", 00:21:06.188 "params": { 00:21:06.188 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.188 "host": "nqn.2016-06.io.spdk:host1", 00:21:06.188 "psk": "key0" 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "nvmf_subsystem_add_ns", 00:21:06.188 "params": { 00:21:06.188 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.188 "namespace": { 00:21:06.188 "nsid": 1, 00:21:06.188 "bdev_name": "malloc0", 00:21:06.188 "nguid": "43318B2296E348C3A987B6E33F7EBD6A", 00:21:06.188 "uuid": "43318b22-96e3-48c3-a987-b6e33f7ebd6a", 00:21:06.188 "no_auto_visible": false 00:21:06.188 } 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "nvmf_subsystem_add_listener", 00:21:06.188 "params": { 00:21:06.188 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.188 "listen_address": { 00:21:06.188 "trtype": "TCP", 00:21:06.188 "adrfam": "IPv4", 00:21:06.188 "traddr": "10.0.0.2", 00:21:06.188 "trsvcid": "4420" 00:21:06.188 }, 00:21:06.188 "secure_channel": false, 00:21:06.188 "sock_impl": "ssl" 00:21:06.188 } 00:21:06.188 } 00:21:06.188 ] 00:21:06.188 } 00:21:06.188 ] 00:21:06.188 }' 00:21:06.188 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:21:06.188 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # bperfcfg='{ 00:21:06.188 "subsystems": [ 00:21:06.188 { 00:21:06.188 "subsystem": "keyring", 00:21:06.188 "config": [ 00:21:06.188 { 00:21:06.188 "method": "keyring_file_add_key", 00:21:06.188 "params": { 00:21:06.188 "name": "key0", 00:21:06.188 "path": "/tmp/tmp.wq2lUTRnNY" 00:21:06.188 } 00:21:06.188 } 00:21:06.188 ] 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "subsystem": "iobuf", 00:21:06.188 "config": [ 00:21:06.188 { 00:21:06.188 "method": "iobuf_set_options", 00:21:06.188 "params": { 00:21:06.188 "small_pool_count": 8192, 00:21:06.188 "large_pool_count": 1024, 00:21:06.188 "small_bufsize": 8192, 00:21:06.188 "large_bufsize": 135168, 00:21:06.188 "enable_numa": false 00:21:06.188 } 00:21:06.188 } 00:21:06.188 ] 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "subsystem": "sock", 00:21:06.188 "config": [ 00:21:06.188 { 00:21:06.188 "method": "sock_set_default_impl", 00:21:06.188 "params": { 00:21:06.188 "impl_name": "posix" 00:21:06.188 } 00:21:06.188 }, 00:21:06.188 { 00:21:06.188 "method": "sock_impl_set_options", 00:21:06.188 "params": { 00:21:06.188 "impl_name": "ssl", 00:21:06.188 "recv_buf_size": 4096, 00:21:06.188 "send_buf_size": 4096, 00:21:06.189 "enable_recv_pipe": true, 00:21:06.189 "enable_quickack": false, 00:21:06.189 "enable_placement_id": 0, 00:21:06.189 "enable_zerocopy_send_server": true, 00:21:06.189 "enable_zerocopy_send_client": false, 00:21:06.189 "zerocopy_threshold": 0, 00:21:06.189 "tls_version": 0, 00:21:06.189 "enable_ktls": false 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "sock_impl_set_options", 00:21:06.189 "params": { 00:21:06.189 "impl_name": "posix", 00:21:06.189 "recv_buf_size": 2097152, 00:21:06.189 "send_buf_size": 2097152, 00:21:06.189 "enable_recv_pipe": true, 00:21:06.189 "enable_quickack": false, 00:21:06.189 "enable_placement_id": 0, 00:21:06.189 "enable_zerocopy_send_server": true, 00:21:06.189 "enable_zerocopy_send_client": false, 00:21:06.189 "zerocopy_threshold": 0, 00:21:06.189 "tls_version": 0, 00:21:06.189 "enable_ktls": false 00:21:06.189 } 00:21:06.189 } 00:21:06.189 ] 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "subsystem": "vmd", 00:21:06.189 "config": [] 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "subsystem": "accel", 00:21:06.189 "config": [ 00:21:06.189 { 00:21:06.189 "method": "accel_set_options", 00:21:06.189 "params": { 00:21:06.189 "small_cache_size": 128, 00:21:06.189 "large_cache_size": 16, 00:21:06.189 "task_count": 2048, 00:21:06.189 "sequence_count": 2048, 00:21:06.189 "buf_count": 2048 00:21:06.189 } 00:21:06.189 } 00:21:06.189 ] 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "subsystem": "bdev", 00:21:06.189 "config": [ 00:21:06.189 { 00:21:06.189 "method": "bdev_set_options", 00:21:06.189 "params": { 00:21:06.189 "bdev_io_pool_size": 65535, 00:21:06.189 "bdev_io_cache_size": 256, 00:21:06.189 "bdev_auto_examine": true, 00:21:06.189 "iobuf_small_cache_size": 128, 00:21:06.189 "iobuf_large_cache_size": 16 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "bdev_raid_set_options", 00:21:06.189 "params": { 00:21:06.189 "process_window_size_kb": 1024, 00:21:06.189 "process_max_bandwidth_mb_sec": 0 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "bdev_iscsi_set_options", 00:21:06.189 "params": { 00:21:06.189 "timeout_sec": 30 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "bdev_nvme_set_options", 00:21:06.189 "params": { 00:21:06.189 "action_on_timeout": "none", 00:21:06.189 "timeout_us": 0, 00:21:06.189 "timeout_admin_us": 0, 00:21:06.189 "keep_alive_timeout_ms": 10000, 00:21:06.189 "arbitration_burst": 0, 00:21:06.189 "low_priority_weight": 0, 00:21:06.189 "medium_priority_weight": 0, 00:21:06.189 "high_priority_weight": 0, 00:21:06.189 "nvme_adminq_poll_period_us": 10000, 00:21:06.189 "nvme_ioq_poll_period_us": 0, 00:21:06.189 "io_queue_requests": 512, 00:21:06.189 "delay_cmd_submit": true, 00:21:06.189 "transport_retry_count": 4, 00:21:06.189 "bdev_retry_count": 3, 00:21:06.189 "transport_ack_timeout": 0, 00:21:06.189 "ctrlr_loss_timeout_sec": 0, 00:21:06.189 "reconnect_delay_sec": 0, 00:21:06.189 "fast_io_fail_timeout_sec": 0, 00:21:06.189 "disable_auto_failback": false, 00:21:06.189 "generate_uuids": false, 00:21:06.189 "transport_tos": 0, 00:21:06.189 "nvme_error_stat": false, 00:21:06.189 "rdma_srq_size": 0, 00:21:06.189 "io_path_stat": false, 00:21:06.189 "allow_accel_sequence": false, 00:21:06.189 "rdma_max_cq_size": 0, 00:21:06.189 "rdma_cm_event_timeout_ms": 0, 00:21:06.189 "dhchap_digests": [ 00:21:06.189 "sha256", 00:21:06.189 "sha384", 00:21:06.189 "sha512" 00:21:06.189 ], 00:21:06.189 "dhchap_dhgroups": [ 00:21:06.189 "null", 00:21:06.189 "ffdhe2048", 00:21:06.189 "ffdhe3072", 00:21:06.189 "ffdhe4096", 00:21:06.189 "ffdhe6144", 00:21:06.189 "ffdhe8192" 00:21:06.189 ] 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "bdev_nvme_attach_controller", 00:21:06.189 "params": { 00:21:06.189 "name": "nvme0", 00:21:06.189 "trtype": "TCP", 00:21:06.189 "adrfam": "IPv4", 00:21:06.189 "traddr": "10.0.0.2", 00:21:06.189 "trsvcid": "4420", 00:21:06.189 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.189 "prchk_reftag": false, 00:21:06.189 "prchk_guard": false, 00:21:06.189 "ctrlr_loss_timeout_sec": 0, 00:21:06.189 "reconnect_delay_sec": 0, 00:21:06.189 "fast_io_fail_timeout_sec": 0, 00:21:06.189 "psk": "key0", 00:21:06.189 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:21:06.189 "hdgst": false, 00:21:06.189 "ddgst": false, 00:21:06.189 "multipath": "multipath" 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "bdev_nvme_set_hotplug", 00:21:06.189 "params": { 00:21:06.189 "period_us": 100000, 00:21:06.189 "enable": false 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "bdev_enable_histogram", 00:21:06.189 "params": { 00:21:06.189 "name": "nvme0n1", 00:21:06.189 "enable": true 00:21:06.189 } 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "method": "bdev_wait_for_examine" 00:21:06.189 } 00:21:06.189 ] 00:21:06.189 }, 00:21:06.189 { 00:21:06.189 "subsystem": "nbd", 00:21:06.189 "config": [] 00:21:06.189 } 00:21:06.189 ] 00:21:06.189 }' 00:21:06.189 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@270 -- # killprocess 3884850 00:21:06.189 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3884850 ']' 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3884850 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3884850 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3884850' 00:21:06.451 killing process with pid 3884850 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3884850 00:21:06.451 Received shutdown signal, test time was about 1.000000 seconds 00:21:06.451 00:21:06.451 Latency(us) 00:21:06.451 [2024-11-26T18:17:19.076Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:06.451 [2024-11-26T18:17:19.076Z] =================================================================================================================== 00:21:06.451 [2024-11-26T18:17:19.076Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3884850 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@271 -- # killprocess 3884784 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3884784 ']' 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3884784 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:06.451 19:17:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3884784 00:21:06.451 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:21:06.451 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:21:06.451 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3884784' 00:21:06.451 killing process with pid 3884784 00:21:06.451 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3884784 00:21:06.451 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3884784 00:21:06.714 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # nvmfappstart -c /dev/fd/62 00:21:06.714 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:21:06.714 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:21:06.714 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # echo '{ 00:21:06.714 "subsystems": [ 00:21:06.714 { 00:21:06.714 "subsystem": "keyring", 00:21:06.714 "config": [ 00:21:06.714 { 00:21:06.714 "method": "keyring_file_add_key", 00:21:06.714 "params": { 00:21:06.714 "name": "key0", 00:21:06.714 "path": "/tmp/tmp.wq2lUTRnNY" 00:21:06.714 } 00:21:06.714 } 00:21:06.714 ] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "iobuf", 00:21:06.714 "config": [ 00:21:06.714 { 00:21:06.714 "method": "iobuf_set_options", 00:21:06.714 "params": { 00:21:06.714 "small_pool_count": 8192, 00:21:06.714 "large_pool_count": 1024, 00:21:06.714 "small_bufsize": 8192, 00:21:06.714 "large_bufsize": 135168, 00:21:06.714 "enable_numa": false 00:21:06.714 } 00:21:06.714 } 00:21:06.714 ] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "sock", 00:21:06.714 "config": [ 00:21:06.714 { 00:21:06.714 "method": "sock_set_default_impl", 00:21:06.714 "params": { 00:21:06.714 "impl_name": "posix" 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "sock_impl_set_options", 00:21:06.714 "params": { 00:21:06.714 "impl_name": "ssl", 00:21:06.714 "recv_buf_size": 4096, 00:21:06.714 "send_buf_size": 4096, 00:21:06.714 "enable_recv_pipe": true, 00:21:06.714 "enable_quickack": false, 00:21:06.714 "enable_placement_id": 0, 00:21:06.714 "enable_zerocopy_send_server": true, 00:21:06.714 "enable_zerocopy_send_client": false, 00:21:06.714 "zerocopy_threshold": 0, 00:21:06.714 "tls_version": 0, 00:21:06.714 "enable_ktls": false 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "sock_impl_set_options", 00:21:06.714 "params": { 00:21:06.714 "impl_name": "posix", 00:21:06.714 "recv_buf_size": 2097152, 00:21:06.714 "send_buf_size": 2097152, 00:21:06.714 "enable_recv_pipe": true, 00:21:06.714 "enable_quickack": false, 00:21:06.714 "enable_placement_id": 0, 00:21:06.714 "enable_zerocopy_send_server": true, 00:21:06.714 "enable_zerocopy_send_client": false, 00:21:06.714 "zerocopy_threshold": 0, 00:21:06.714 "tls_version": 0, 00:21:06.714 "enable_ktls": false 00:21:06.714 } 00:21:06.714 } 00:21:06.714 ] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "vmd", 00:21:06.714 "config": [] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "accel", 00:21:06.714 "config": [ 00:21:06.714 { 00:21:06.714 "method": "accel_set_options", 00:21:06.714 "params": { 00:21:06.714 "small_cache_size": 128, 00:21:06.714 "large_cache_size": 16, 00:21:06.714 "task_count": 2048, 00:21:06.714 "sequence_count": 2048, 00:21:06.714 "buf_count": 2048 00:21:06.714 } 00:21:06.714 } 00:21:06.714 ] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "bdev", 00:21:06.714 "config": [ 00:21:06.714 { 00:21:06.714 "method": "bdev_set_options", 00:21:06.714 "params": { 00:21:06.714 "bdev_io_pool_size": 65535, 00:21:06.714 "bdev_io_cache_size": 256, 00:21:06.714 "bdev_auto_examine": true, 00:21:06.714 "iobuf_small_cache_size": 128, 00:21:06.714 "iobuf_large_cache_size": 16 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "bdev_raid_set_options", 00:21:06.714 "params": { 00:21:06.714 "process_window_size_kb": 1024, 00:21:06.714 "process_max_bandwidth_mb_sec": 0 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "bdev_iscsi_set_options", 00:21:06.714 "params": { 00:21:06.714 "timeout_sec": 30 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "bdev_nvme_set_options", 00:21:06.714 "params": { 00:21:06.714 "action_on_timeout": "none", 00:21:06.714 "timeout_us": 0, 00:21:06.714 "timeout_admin_us": 0, 00:21:06.714 "keep_alive_timeout_ms": 10000, 00:21:06.714 "arbitration_burst": 0, 00:21:06.714 "low_priority_weight": 0, 00:21:06.714 "medium_priority_weight": 0, 00:21:06.714 "high_priority_weight": 0, 00:21:06.714 "nvme_adminq_poll_period_us": 10000, 00:21:06.714 "nvme_ioq_poll_period_us": 0, 00:21:06.714 "io_queue_requests": 0, 00:21:06.714 "delay_cmd_submit": true, 00:21:06.714 "transport_retry_count": 4, 00:21:06.714 "bdev_retry_count": 3, 00:21:06.714 "transport_ack_timeout": 0, 00:21:06.714 "ctrlr_loss_timeout_sec": 0, 00:21:06.714 "reconnect_delay_sec": 0, 00:21:06.714 "fast_io_fail_timeout_sec": 0, 00:21:06.714 "disable_auto_failback": false, 00:21:06.714 "generate_uuids": false, 00:21:06.714 "transport_tos": 0, 00:21:06.714 "nvme_error_stat": false, 00:21:06.714 "rdma_srq_size": 0, 00:21:06.714 "io_path_stat": false, 00:21:06.714 "allow_accel_sequence": false, 00:21:06.714 "rdma_max_cq_size": 0, 00:21:06.714 "rdma_cm_event_timeout_ms": 0, 00:21:06.714 "dhchap_digests": [ 00:21:06.714 "sha256", 00:21:06.714 "sha384", 00:21:06.714 "sha512" 00:21:06.714 ], 00:21:06.714 "dhchap_dhgroups": [ 00:21:06.714 "null", 00:21:06.714 "ffdhe2048", 00:21:06.714 "ffdhe3072", 00:21:06.714 "ffdhe4096", 00:21:06.714 "ffdhe6144", 00:21:06.714 "ffdhe8192" 00:21:06.714 ] 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "bdev_nvme_set_hotplug", 00:21:06.714 "params": { 00:21:06.714 "period_us": 100000, 00:21:06.714 "enable": false 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "bdev_malloc_create", 00:21:06.714 "params": { 00:21:06.714 "name": "malloc0", 00:21:06.714 "num_blocks": 8192, 00:21:06.714 "block_size": 4096, 00:21:06.714 "physical_block_size": 4096, 00:21:06.714 "uuid": "43318b22-96e3-48c3-a987-b6e33f7ebd6a", 00:21:06.714 "optimal_io_boundary": 0, 00:21:06.714 "md_size": 0, 00:21:06.714 "dif_type": 0, 00:21:06.714 "dif_is_head_of_md": false, 00:21:06.714 "dif_pi_format": 0 00:21:06.714 } 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "method": "bdev_wait_for_examine" 00:21:06.714 } 00:21:06.714 ] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "nbd", 00:21:06.714 "config": [] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "scheduler", 00:21:06.714 "config": [ 00:21:06.714 { 00:21:06.714 "method": "framework_set_scheduler", 00:21:06.714 "params": { 00:21:06.714 "name": "static" 00:21:06.714 } 00:21:06.714 } 00:21:06.714 ] 00:21:06.714 }, 00:21:06.714 { 00:21:06.714 "subsystem": "nvmf", 00:21:06.714 "config": [ 00:21:06.714 { 00:21:06.714 "method": "nvmf_set_config", 00:21:06.714 "params": { 00:21:06.714 "discovery_filter": "match_any", 00:21:06.714 "admin_cmd_passthru": { 00:21:06.714 "identify_ctrlr": false 00:21:06.714 }, 00:21:06.715 "dhchap_digests": [ 00:21:06.715 "sha256", 00:21:06.715 "sha384", 00:21:06.715 "sha512" 00:21:06.715 ], 00:21:06.715 "dhchap_dhgroups": [ 00:21:06.715 "null", 00:21:06.715 "ffdhe2048", 00:21:06.715 "ffdhe3072", 00:21:06.715 "ffdhe4096", 00:21:06.715 "ffdhe6144", 00:21:06.715 "ffdhe8192" 00:21:06.715 ] 00:21:06.715 } 00:21:06.715 }, 00:21:06.715 { 00:21:06.715 "method": "nvmf_set_max_subsystems", 00:21:06.715 "params": { 00:21:06.715 "max_subsystems": 1024 00:21:06.715 } 00:21:06.715 }, 00:21:06.715 { 00:21:06.715 "method": "nvmf_set_crdt", 00:21:06.715 "params": { 00:21:06.715 "crdt1": 0, 00:21:06.715 "crdt2": 0, 00:21:06.715 "crdt3": 0 00:21:06.715 } 00:21:06.715 }, 00:21:06.715 { 00:21:06.715 "method": "nvmf_create_transport", 00:21:06.715 "params": { 00:21:06.715 "trtype": "TCP", 00:21:06.715 "max_queue_depth": 128, 00:21:06.715 "max_io_qpairs_per_ctrlr": 127, 00:21:06.715 "in_capsule_data_size": 4096, 00:21:06.715 "max_io_size": 131072, 00:21:06.715 "io_unit_size": 131072, 00:21:06.715 "max_aq_depth": 128, 00:21:06.715 "num_shared_buffers": 511, 00:21:06.715 "buf_cache_size": 4294967295, 00:21:06.715 "dif_insert_or_strip": false, 00:21:06.715 "zcopy": false, 00:21:06.715 "c2h_success": false, 00:21:06.715 "sock_priority": 0, 00:21:06.715 "abort_timeout_sec": 1, 00:21:06.715 "ack_timeout": 0, 00:21:06.715 "data_wr_pool_size": 0 00:21:06.715 } 00:21:06.715 }, 00:21:06.715 { 00:21:06.715 "method": "nvmf_create_subsystem", 00:21:06.715 "params": { 00:21:06.715 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.715 "allow_any_host": false, 00:21:06.715 "serial_number": "00000000000000000000", 00:21:06.715 "model_number": "SPDK bdev Controller", 00:21:06.715 "max_namespaces": 32, 00:21:06.715 "min_cntlid": 1, 00:21:06.715 "max_cntlid": 65519, 00:21:06.715 "ana_reporting": false 00:21:06.715 } 00:21:06.715 }, 00:21:06.715 { 00:21:06.715 "method": "nvmf_subsystem_add_host", 00:21:06.715 "params": { 00:21:06.715 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.715 "host": "nqn.2016-06.io.spdk:host1", 00:21:06.715 "psk": "key0" 00:21:06.715 } 00:21:06.715 }, 00:21:06.715 { 00:21:06.715 "method": "nvmf_subsystem_add_ns", 00:21:06.715 "params": { 00:21:06.715 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.715 "namespace": { 00:21:06.715 "nsid": 1, 00:21:06.715 "bdev_name": "malloc0", 00:21:06.715 "nguid": "43318B2296E348C3A987B6E33F7EBD6A", 00:21:06.715 "uuid": "43318b22-96e3-48c3-a987-b6e33f7ebd6a", 00:21:06.715 "no_auto_visible": false 00:21:06.715 } 00:21:06.715 } 00:21:06.715 }, 00:21:06.715 { 00:21:06.715 "method": "nvmf_subsystem_add_listener", 00:21:06.715 "params": { 00:21:06.715 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.715 "listen_address": { 00:21:06.715 "trtype": "TCP", 00:21:06.715 "adrfam": "IPv4", 00:21:06.715 "traddr": "10.0.0.2", 00:21:06.715 "trsvcid": "4420" 00:21:06.715 }, 00:21:06.715 "secure_channel": false, 00:21:06.715 "sock_impl": "ssl" 00:21:06.715 } 00:21:06.715 } 00:21:06.715 ] 00:21:06.715 } 00:21:06.715 ] 00:21:06.715 }' 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=3885499 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 3885499 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -c /dev/fd/62 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3885499 ']' 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:06.715 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:06.715 19:17:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:06.715 [2024-11-26 19:17:19.231194] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:06.715 [2024-11-26 19:17:19.231246] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:06.715 [2024-11-26 19:17:19.316657] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:06.976 [2024-11-26 19:17:19.349982] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:06.976 [2024-11-26 19:17:19.350016] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:06.976 [2024-11-26 19:17:19.350024] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:06.976 [2024-11-26 19:17:19.350031] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:06.976 [2024-11-26 19:17:19.350036] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:06.976 [2024-11-26 19:17:19.350643] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:06.976 [2024-11-26 19:17:19.551010] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:06.976 [2024-11-26 19:17:19.583018] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:21:06.976 [2024-11-26 19:17:19.583236] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@276 -- # bdevperf_pid=3885841 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@277 -- # waitforlisten 3885841 /var/tmp/bdevperf.sock 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 3885841 ']' 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:07.548 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 -c /dev/fd/63 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:07.548 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # echo '{ 00:21:07.548 "subsystems": [ 00:21:07.548 { 00:21:07.548 "subsystem": "keyring", 00:21:07.548 "config": [ 00:21:07.548 { 00:21:07.548 "method": "keyring_file_add_key", 00:21:07.548 "params": { 00:21:07.548 "name": "key0", 00:21:07.548 "path": "/tmp/tmp.wq2lUTRnNY" 00:21:07.548 } 00:21:07.548 } 00:21:07.548 ] 00:21:07.548 }, 00:21:07.548 { 00:21:07.548 "subsystem": "iobuf", 00:21:07.548 "config": [ 00:21:07.548 { 00:21:07.548 "method": "iobuf_set_options", 00:21:07.548 "params": { 00:21:07.548 "small_pool_count": 8192, 00:21:07.548 "large_pool_count": 1024, 00:21:07.548 "small_bufsize": 8192, 00:21:07.548 "large_bufsize": 135168, 00:21:07.548 "enable_numa": false 00:21:07.548 } 00:21:07.548 } 00:21:07.548 ] 00:21:07.548 }, 00:21:07.548 { 00:21:07.548 "subsystem": "sock", 00:21:07.548 "config": [ 00:21:07.548 { 00:21:07.548 "method": "sock_set_default_impl", 00:21:07.548 "params": { 00:21:07.548 "impl_name": "posix" 00:21:07.548 } 00:21:07.548 }, 00:21:07.548 { 00:21:07.548 "method": "sock_impl_set_options", 00:21:07.548 "params": { 00:21:07.548 "impl_name": "ssl", 00:21:07.548 "recv_buf_size": 4096, 00:21:07.549 "send_buf_size": 4096, 00:21:07.549 "enable_recv_pipe": true, 00:21:07.549 "enable_quickack": false, 00:21:07.549 "enable_placement_id": 0, 00:21:07.549 "enable_zerocopy_send_server": true, 00:21:07.549 "enable_zerocopy_send_client": false, 00:21:07.549 "zerocopy_threshold": 0, 00:21:07.549 "tls_version": 0, 00:21:07.549 "enable_ktls": false 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "sock_impl_set_options", 00:21:07.549 "params": { 00:21:07.549 "impl_name": "posix", 00:21:07.549 "recv_buf_size": 2097152, 00:21:07.549 "send_buf_size": 2097152, 00:21:07.549 "enable_recv_pipe": true, 00:21:07.549 "enable_quickack": false, 00:21:07.549 "enable_placement_id": 0, 00:21:07.549 "enable_zerocopy_send_server": true, 00:21:07.549 "enable_zerocopy_send_client": false, 00:21:07.549 "zerocopy_threshold": 0, 00:21:07.549 "tls_version": 0, 00:21:07.549 "enable_ktls": false 00:21:07.549 } 00:21:07.549 } 00:21:07.549 ] 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "subsystem": "vmd", 00:21:07.549 "config": [] 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "subsystem": "accel", 00:21:07.549 "config": [ 00:21:07.549 { 00:21:07.549 "method": "accel_set_options", 00:21:07.549 "params": { 00:21:07.549 "small_cache_size": 128, 00:21:07.549 "large_cache_size": 16, 00:21:07.549 "task_count": 2048, 00:21:07.549 "sequence_count": 2048, 00:21:07.549 "buf_count": 2048 00:21:07.549 } 00:21:07.549 } 00:21:07.549 ] 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "subsystem": "bdev", 00:21:07.549 "config": [ 00:21:07.549 { 00:21:07.549 "method": "bdev_set_options", 00:21:07.549 "params": { 00:21:07.549 "bdev_io_pool_size": 65535, 00:21:07.549 "bdev_io_cache_size": 256, 00:21:07.549 "bdev_auto_examine": true, 00:21:07.549 "iobuf_small_cache_size": 128, 00:21:07.549 "iobuf_large_cache_size": 16 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "bdev_raid_set_options", 00:21:07.549 "params": { 00:21:07.549 "process_window_size_kb": 1024, 00:21:07.549 "process_max_bandwidth_mb_sec": 0 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "bdev_iscsi_set_options", 00:21:07.549 "params": { 00:21:07.549 "timeout_sec": 30 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "bdev_nvme_set_options", 00:21:07.549 "params": { 00:21:07.549 "action_on_timeout": "none", 00:21:07.549 "timeout_us": 0, 00:21:07.549 "timeout_admin_us": 0, 00:21:07.549 "keep_alive_timeout_ms": 10000, 00:21:07.549 "arbitration_burst": 0, 00:21:07.549 "low_priority_weight": 0, 00:21:07.549 "medium_priority_weight": 0, 00:21:07.549 "high_priority_weight": 0, 00:21:07.549 "nvme_adminq_poll_period_us": 10000, 00:21:07.549 "nvme_ioq_poll_period_us": 0, 00:21:07.549 "io_queue_requests": 512, 00:21:07.549 "delay_cmd_submit": true, 00:21:07.549 "transport_retry_count": 4, 00:21:07.549 "bdev_retry_count": 3, 00:21:07.549 "transport_ack_timeout": 0, 00:21:07.549 "ctrlr_loss_timeout_sec": 0, 00:21:07.549 "reconnect_delay_sec": 0, 00:21:07.549 "fast_io_fail_timeout_sec": 0, 00:21:07.549 "disable_auto_failback": false, 00:21:07.549 "generate_uuids": false, 00:21:07.549 "transport_tos": 0, 00:21:07.549 "nvme_error_stat": false, 00:21:07.549 "rdma_srq_size": 0, 00:21:07.549 "io_path_stat": false, 00:21:07.549 "allow_accel_sequence": false, 00:21:07.549 "rdma_max_cq_size": 0, 00:21:07.549 "rdma_cm_event_timeout_ms": 0, 00:21:07.549 "dhchap_digests": [ 00:21:07.549 "sha256", 00:21:07.549 "sha384", 00:21:07.549 "sha512" 00:21:07.549 ], 00:21:07.549 "dhchap_dhgroups": [ 00:21:07.549 "null", 00:21:07.549 "ffdhe2048", 00:21:07.549 "ffdhe3072", 00:21:07.549 "ffdhe4096", 00:21:07.549 "ffdhe6144", 00:21:07.549 "ffdhe8192" 00:21:07.549 ] 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "bdev_nvme_attach_controller", 00:21:07.549 "params": { 00:21:07.549 "name": "nvme0", 00:21:07.549 "trtype": "TCP", 00:21:07.549 "adrfam": "IPv4", 00:21:07.549 "traddr": "10.0.0.2", 00:21:07.549 "trsvcid": "4420", 00:21:07.549 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:07.549 "prchk_reftag": false, 00:21:07.549 "prchk_guard": false, 00:21:07.549 "ctrlr_loss_timeout_sec": 0, 00:21:07.549 "reconnect_delay_sec": 0, 00:21:07.549 "fast_io_fail_timeout_sec": 0, 00:21:07.549 "psk": "key0", 00:21:07.549 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:21:07.549 "hdgst": false, 00:21:07.549 "ddgst": false, 00:21:07.549 "multipath": "multipath" 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "bdev_nvme_set_hotplug", 00:21:07.549 "params": { 00:21:07.549 "period_us": 100000, 00:21:07.549 "enable": false 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "bdev_enable_histogram", 00:21:07.549 "params": { 00:21:07.549 "name": "nvme0n1", 00:21:07.549 "enable": true 00:21:07.549 } 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "method": "bdev_wait_for_examine" 00:21:07.549 } 00:21:07.549 ] 00:21:07.549 }, 00:21:07.549 { 00:21:07.549 "subsystem": "nbd", 00:21:07.549 "config": [] 00:21:07.549 } 00:21:07.549 ] 00:21:07.549 }' 00:21:07.549 [2024-11-26 19:17:20.134417] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:07.549 [2024-11-26 19:17:20.134473] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3885841 ] 00:21:07.810 [2024-11-26 19:17:20.223656] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:07.810 [2024-11-26 19:17:20.253576] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:07.810 [2024-11-26 19:17:20.389791] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:21:08.381 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:08.381 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:21:08.381 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:21:08.381 19:17:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # jq -r '.[].name' 00:21:08.641 19:17:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:08.641 19:17:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:21:08.641 Running I/O for 1 seconds... 00:21:09.838 5259.00 IOPS, 20.54 MiB/s 00:21:09.838 Latency(us) 00:21:09.838 [2024-11-26T18:17:22.463Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:09.838 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:21:09.838 Verification LBA range: start 0x0 length 0x2000 00:21:09.838 nvme0n1 : 1.02 5283.49 20.64 0.00 0.00 23994.69 6307.84 28617.39 00:21:09.838 [2024-11-26T18:17:22.463Z] =================================================================================================================== 00:21:09.838 [2024-11-26T18:17:22.463Z] Total : 5283.49 20.64 0.00 0.00 23994.69 6307.84 28617.39 00:21:09.838 { 00:21:09.838 "results": [ 00:21:09.838 { 00:21:09.838 "job": "nvme0n1", 00:21:09.838 "core_mask": "0x2", 00:21:09.838 "workload": "verify", 00:21:09.838 "status": "finished", 00:21:09.838 "verify_range": { 00:21:09.838 "start": 0, 00:21:09.838 "length": 8192 00:21:09.838 }, 00:21:09.838 "queue_depth": 128, 00:21:09.838 "io_size": 4096, 00:21:09.838 "runtime": 1.01978, 00:21:09.838 "iops": 5283.492517994077, 00:21:09.838 "mibps": 20.638642648414365, 00:21:09.838 "io_failed": 0, 00:21:09.838 "io_timeout": 0, 00:21:09.838 "avg_latency_us": 23994.69184855234, 00:21:09.838 "min_latency_us": 6307.84, 00:21:09.838 "max_latency_us": 28617.386666666665 00:21:09.838 } 00:21:09.838 ], 00:21:09.838 "core_count": 1 00:21:09.838 } 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@282 -- # trap - SIGINT SIGTERM EXIT 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@283 -- # cleanup 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@15 -- # process_shm --id 0 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@812 -- # type=--id 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@813 -- # id=0 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@824 -- # for n in $shm_files 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:21:09.838 nvmf_trace.0 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@827 -- # return 0 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@16 -- # killprocess 3885841 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3885841 ']' 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3885841 00:21:09.838 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:21:09.839 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:09.839 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3885841 00:21:09.839 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:21:09.839 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:21:09.839 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3885841' 00:21:09.839 killing process with pid 3885841 00:21:09.839 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3885841 00:21:09.839 Received shutdown signal, test time was about 1.000000 seconds 00:21:09.839 00:21:09.839 Latency(us) 00:21:09.839 [2024-11-26T18:17:22.464Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:09.839 [2024-11-26T18:17:22.464Z] =================================================================================================================== 00:21:09.839 [2024-11-26T18:17:22.464Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:21:09.839 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3885841 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@17 -- # nvmftestfini 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@516 -- # nvmfcleanup 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@121 -- # sync 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@124 -- # set +e 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@125 -- # for i in {1..20} 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:21:10.099 rmmod nvme_tcp 00:21:10.099 rmmod nvme_fabrics 00:21:10.099 rmmod nvme_keyring 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@128 -- # set -e 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@129 -- # return 0 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@517 -- # '[' -n 3885499 ']' 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@518 -- # killprocess 3885499 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 3885499 ']' 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 3885499 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3885499 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3885499' 00:21:10.099 killing process with pid 3885499 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 3885499 00:21:10.099 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 3885499 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@297 -- # iptr 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@791 -- # iptables-save 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@791 -- # iptables-restore 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@302 -- # remove_spdk_ns 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:10.359 19:17:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@18 -- # rm -f /tmp/tmp.C8WQVOIq90 /tmp/tmp.KAnagrqvyX /tmp/tmp.wq2lUTRnNY 00:21:12.273 00:21:12.273 real 1m24.160s 00:21:12.273 user 2m8.488s 00:21:12.273 sys 0m27.754s 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1130 -- # xtrace_disable 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:21:12.273 ************************************ 00:21:12.273 END TEST nvmf_tls 00:21:12.273 ************************************ 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@42 -- # run_test nvmf_fips /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:21:12.273 19:17:24 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:21:12.534 ************************************ 00:21:12.534 START TEST nvmf_fips 00:21:12.534 ************************************ 00:21:12.534 19:17:24 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:21:12.534 * Looking for test storage... 00:21:12.534 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips 00:21:12.534 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:21:12.534 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1693 -- # lcov --version 00:21:12.534 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:21:12.534 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:21:12.534 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:21:12.534 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:21:12.534 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=<' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=2 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@345 -- # : 1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 2 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=2 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 2 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=2 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # return 0 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:21:12.535 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:12.535 --rc genhtml_branch_coverage=1 00:21:12.535 --rc genhtml_function_coverage=1 00:21:12.535 --rc genhtml_legend=1 00:21:12.535 --rc geninfo_all_blocks=1 00:21:12.535 --rc geninfo_unexecuted_blocks=1 00:21:12.535 00:21:12.535 ' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:21:12.535 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:12.535 --rc genhtml_branch_coverage=1 00:21:12.535 --rc genhtml_function_coverage=1 00:21:12.535 --rc genhtml_legend=1 00:21:12.535 --rc geninfo_all_blocks=1 00:21:12.535 --rc geninfo_unexecuted_blocks=1 00:21:12.535 00:21:12.535 ' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:21:12.535 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:12.535 --rc genhtml_branch_coverage=1 00:21:12.535 --rc genhtml_function_coverage=1 00:21:12.535 --rc genhtml_legend=1 00:21:12.535 --rc geninfo_all_blocks=1 00:21:12.535 --rc geninfo_unexecuted_blocks=1 00:21:12.535 00:21:12.535 ' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:21:12.535 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:12.535 --rc genhtml_branch_coverage=1 00:21:12.535 --rc genhtml_function_coverage=1 00:21:12.535 --rc genhtml_legend=1 00:21:12.535 --rc geninfo_all_blocks=1 00:21:12.535 --rc geninfo_unexecuted_blocks=1 00:21:12.535 00:21:12.535 ' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # uname -s 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@15 -- # shopt -s extglob 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@5 -- # export PATH 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@51 -- # : 0 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:21:12.535 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@55 -- # have_pci_nics=0 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@90 -- # check_openssl_version 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@84 -- # local target=3.0.0 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # openssl version 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # awk '{print $2}' 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # ge 3.1.1 3.0.0 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@376 -- # cmp_versions 3.1.1 '>=' 3.0.0 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:21:12.535 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=>=' 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@348 -- # : 1 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=3 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v++ )) 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:21:12.536 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 0 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=0 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 0 =~ ^[0-9]+$ ]] 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 0 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=0 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # return 0 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # openssl info -modulesdir 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # [[ ! -f /usr/lib64/ossl-modules/fips.so ]] 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # openssl fipsinstall -help 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # warn='This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode' 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@102 -- # [[ This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode == \T\h\i\s\ \c\o\m\m\a\n\d\ \i\s\ \n\o\t\ \e\n\a\b\l\e\d* ]] 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # export callback=build_openssl_config 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # callback=build_openssl_config 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@114 -- # build_openssl_config 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@38 -- # cat 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@58 -- # [[ ! -t 0 ]] 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@59 -- # cat - 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # export OPENSSL_CONF=spdk_fips.conf 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # OPENSSL_CONF=spdk_fips.conf 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # mapfile -t providers 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # openssl list -providers 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # grep name 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # (( 2 != 2 )) 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: openssl base provider != *base* ]] 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: red hat enterprise linux 9 - openssl fips provider != *fips* ]] 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # NOT openssl md5 /dev/fd/62 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@652 -- # local es=0 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@654 -- # valid_exec_arg openssl md5 /dev/fd/62 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@640 -- # local arg=openssl 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # : 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # type -t openssl 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@646 -- # type -P openssl 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:21:12.797 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@646 -- # arg=/usr/bin/openssl 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@646 -- # [[ -x /usr/bin/openssl ]] 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@655 -- # openssl md5 /dev/fd/62 00:21:12.798 Error setting digest 00:21:12.798 4082922FA97F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:crypto/evp/evp_fetch.c:341:Global default library context, Algorithm (MD5 : 95), Properties () 00:21:12.798 4082922FA97F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:crypto/evp/digest.c:272: 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@655 -- # es=1 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@131 -- # nvmftestinit 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@476 -- # prepare_net_devs 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@438 -- # local -g is_hw=no 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # remove_spdk_ns 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@309 -- # xtrace_disable 00:21:12.798 19:17:25 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # pci_devs=() 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # local -a pci_devs 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # pci_net_devs=() 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # pci_drivers=() 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # local -A pci_drivers 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # net_devs=() 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # local -ga net_devs 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # e810=() 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # local -ga e810 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # x722=() 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # local -ga x722 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # mlx=() 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # local -ga mlx 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:21:20.943 Found 0000:31:00.0 (0x8086 - 0x159b) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:21:20.943 Found 0000:31:00.1 (0x8086 - 0x159b) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@418 -- # [[ up == up ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:21:20.943 Found net devices under 0000:31:00.0: cvl_0_0 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@418 -- # [[ up == up ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:21:20.943 Found net devices under 0000:31:00.1: cvl_0_1 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # is_hw=yes 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:21:20.943 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:21:20.944 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:21:20.944 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:20.944 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:20.944 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:20.944 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:21:20.944 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:21:21.208 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:21.208 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.508 ms 00:21:21.208 00:21:21.208 --- 10.0.0.2 ping statistics --- 00:21:21.208 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:21.208 rtt min/avg/max/mdev = 0.508/0.508/0.508/0.000 ms 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:21.208 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:21.208 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.322 ms 00:21:21.208 00:21:21.208 --- 10.0.0.1 ping statistics --- 00:21:21.208 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:21.208 rtt min/avg/max/mdev = 0.322/0.322/0.322/0.000 ms 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@450 -- # return 0 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:21:21.208 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@132 -- # nvmfappstart -m 0x2 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@726 -- # xtrace_disable 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@509 -- # nvmfpid=3890982 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@510 -- # waitforlisten 3890982 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # '[' -z 3890982 ']' 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:21.209 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:21.209 19:17:33 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:21:21.209 [2024-11-26 19:17:33.779817] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:21.209 [2024-11-26 19:17:33.779898] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:21.470 [2024-11-26 19:17:33.889713] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:21.470 [2024-11-26 19:17:33.939692] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:21.470 [2024-11-26 19:17:33.939750] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:21.470 [2024-11-26 19:17:33.939759] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:21.470 [2024-11-26 19:17:33.939766] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:21.470 [2024-11-26 19:17:33.939772] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:21.470 [2024-11-26 19:17:33.940632] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@868 -- # return 0 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@732 -- # xtrace_disable 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@134 -- # trap cleanup EXIT 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@137 -- # key=NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # mktemp -t spdk-psk.XXX 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # key_path=/tmp/spdk-psk.GH9 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@139 -- # echo -n NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@140 -- # chmod 0600 /tmp/spdk-psk.GH9 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@142 -- # setup_nvmf_tgt_conf /tmp/spdk-psk.GH9 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@22 -- # local key=/tmp/spdk-psk.GH9 00:21:22.041 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:21:22.303 [2024-11-26 19:17:34.783564] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:22.303 [2024-11-26 19:17:34.799553] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:21:22.303 [2024-11-26 19:17:34.799927] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:22.303 malloc0 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@145 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@148 -- # bdevperf_pid=3891266 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@149 -- # waitforlisten 3891266 /var/tmp/bdevperf.sock 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@146 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # '[' -z 3891266 ']' 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:22.303 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:22.303 19:17:34 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:21:22.564 [2024-11-26 19:17:34.939617] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:22.564 [2024-11-26 19:17:34.939685] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3891266 ] 00:21:22.564 [2024-11-26 19:17:35.011679] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:22.564 [2024-11-26 19:17:35.047652] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:21:23.134 19:17:35 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:23.134 19:17:35 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@868 -- # return 0 00:21:23.134 19:17:35 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/spdk-psk.GH9 00:21:23.393 19:17:35 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@152 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:21:23.654 [2024-11-26 19:17:36.067338] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:21:23.654 TLSTESTn1 00:21:23.654 19:17:36 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@156 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:21:23.654 Running I/O for 10 seconds... 00:21:25.981 4609.00 IOPS, 18.00 MiB/s [2024-11-26T18:17:39.547Z] 5569.50 IOPS, 21.76 MiB/s [2024-11-26T18:17:40.488Z] 5627.33 IOPS, 21.98 MiB/s [2024-11-26T18:17:41.429Z] 5666.25 IOPS, 22.13 MiB/s [2024-11-26T18:17:42.369Z] 5822.40 IOPS, 22.74 MiB/s [2024-11-26T18:17:43.309Z] 5756.17 IOPS, 22.49 MiB/s [2024-11-26T18:17:44.696Z] 5791.71 IOPS, 22.62 MiB/s [2024-11-26T18:17:45.637Z] 5831.50 IOPS, 22.78 MiB/s [2024-11-26T18:17:46.581Z] 5901.78 IOPS, 23.05 MiB/s [2024-11-26T18:17:46.581Z] 5885.60 IOPS, 22.99 MiB/s 00:21:33.956 Latency(us) 00:21:33.956 [2024-11-26T18:17:46.581Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:33.956 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:21:33.956 Verification LBA range: start 0x0 length 0x2000 00:21:33.956 TLSTESTn1 : 10.01 5890.87 23.01 0.00 0.00 21696.52 5324.80 96556.37 00:21:33.956 [2024-11-26T18:17:46.581Z] =================================================================================================================== 00:21:33.956 [2024-11-26T18:17:46.581Z] Total : 5890.87 23.01 0.00 0.00 21696.52 5324.80 96556.37 00:21:33.956 { 00:21:33.956 "results": [ 00:21:33.956 { 00:21:33.956 "job": "TLSTESTn1", 00:21:33.956 "core_mask": "0x4", 00:21:33.956 "workload": "verify", 00:21:33.956 "status": "finished", 00:21:33.956 "verify_range": { 00:21:33.956 "start": 0, 00:21:33.956 "length": 8192 00:21:33.956 }, 00:21:33.956 "queue_depth": 128, 00:21:33.956 "io_size": 4096, 00:21:33.956 "runtime": 10.012606, 00:21:33.956 "iops": 5890.873964280628, 00:21:33.956 "mibps": 23.011226422971202, 00:21:33.956 "io_failed": 0, 00:21:33.956 "io_timeout": 0, 00:21:33.956 "avg_latency_us": 21696.515402743164, 00:21:33.956 "min_latency_us": 5324.8, 00:21:33.956 "max_latency_us": 96556.37333333334 00:21:33.956 } 00:21:33.956 ], 00:21:33.956 "core_count": 1 00:21:33.956 } 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@1 -- # cleanup 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@15 -- # process_shm --id 0 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@812 -- # type=--id 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@813 -- # id=0 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@824 -- # for n in $shm_files 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:21:33.956 nvmf_trace.0 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@827 -- # return 0 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@16 -- # killprocess 3891266 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # '[' -z 3891266 ']' 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@958 -- # kill -0 3891266 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # uname 00:21:33.956 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3891266 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3891266' 00:21:33.957 killing process with pid 3891266 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@973 -- # kill 3891266 00:21:33.957 Received shutdown signal, test time was about 10.000000 seconds 00:21:33.957 00:21:33.957 Latency(us) 00:21:33.957 [2024-11-26T18:17:46.582Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:33.957 [2024-11-26T18:17:46.582Z] =================================================================================================================== 00:21:33.957 [2024-11-26T18:17:46.582Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@978 -- # wait 3891266 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@17 -- # nvmftestfini 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@516 -- # nvmfcleanup 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@121 -- # sync 00:21:33.957 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@124 -- # set +e 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@125 -- # for i in {1..20} 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:21:34.218 rmmod nvme_tcp 00:21:34.218 rmmod nvme_fabrics 00:21:34.218 rmmod nvme_keyring 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@128 -- # set -e 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@129 -- # return 0 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@517 -- # '[' -n 3890982 ']' 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@518 -- # killprocess 3890982 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # '[' -z 3890982 ']' 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@958 -- # kill -0 3890982 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # uname 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3890982 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3890982' 00:21:34.218 killing process with pid 3890982 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@973 -- # kill 3890982 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@978 -- # wait 3890982 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:21:34.218 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:21:34.219 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@297 -- # iptr 00:21:34.219 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@791 -- # iptables-save 00:21:34.219 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@791 -- # iptables-restore 00:21:34.219 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:21:34.478 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:21:34.478 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@302 -- # remove_spdk_ns 00:21:34.478 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:34.478 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:34.478 19:17:46 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@18 -- # rm -f /tmp/spdk-psk.GH9 00:21:36.389 00:21:36.389 real 0m24.017s 00:21:36.389 user 0m24.991s 00:21:36.389 sys 0m10.285s 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1130 -- # xtrace_disable 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:21:36.389 ************************************ 00:21:36.389 END TEST nvmf_fips 00:21:36.389 ************************************ 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@43 -- # run_test nvmf_control_msg_list /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:21:36.389 19:17:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:21:36.389 ************************************ 00:21:36.389 START TEST nvmf_control_msg_list 00:21:36.389 ************************************ 00:21:36.389 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:21:36.651 * Looking for test storage... 00:21:36.651 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1693 -- # lcov --version 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@333 -- # local ver1 ver1_l 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@334 -- # local ver2 ver2_l 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # IFS=.-: 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # read -ra ver1 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # IFS=.-: 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # read -ra ver2 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@338 -- # local 'op=<' 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@340 -- # ver1_l=2 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@341 -- # ver2_l=1 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@344 -- # case "$op" in 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@345 -- # : 1 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v = 0 )) 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # decimal 1 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=1 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 1 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # ver1[v]=1 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # decimal 2 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=2 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 2 00:21:36.651 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # ver2[v]=2 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # return 0 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:21:36.652 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:36.652 --rc genhtml_branch_coverage=1 00:21:36.652 --rc genhtml_function_coverage=1 00:21:36.652 --rc genhtml_legend=1 00:21:36.652 --rc geninfo_all_blocks=1 00:21:36.652 --rc geninfo_unexecuted_blocks=1 00:21:36.652 00:21:36.652 ' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:21:36.652 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:36.652 --rc genhtml_branch_coverage=1 00:21:36.652 --rc genhtml_function_coverage=1 00:21:36.652 --rc genhtml_legend=1 00:21:36.652 --rc geninfo_all_blocks=1 00:21:36.652 --rc geninfo_unexecuted_blocks=1 00:21:36.652 00:21:36.652 ' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:21:36.652 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:36.652 --rc genhtml_branch_coverage=1 00:21:36.652 --rc genhtml_function_coverage=1 00:21:36.652 --rc genhtml_legend=1 00:21:36.652 --rc geninfo_all_blocks=1 00:21:36.652 --rc geninfo_unexecuted_blocks=1 00:21:36.652 00:21:36.652 ' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:21:36.652 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:36.652 --rc genhtml_branch_coverage=1 00:21:36.652 --rc genhtml_function_coverage=1 00:21:36.652 --rc genhtml_legend=1 00:21:36.652 --rc geninfo_all_blocks=1 00:21:36.652 --rc geninfo_unexecuted_blocks=1 00:21:36.652 00:21:36.652 ' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # uname -s 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@15 -- # shopt -s extglob 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@5 -- # export PATH 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@51 -- # : 0 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:21:36.652 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@55 -- # have_pci_nics=0 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@12 -- # nvmftestinit 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@476 -- # prepare_net_devs 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@438 -- # local -g is_hw=no 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # remove_spdk_ns 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:36.652 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:36.653 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:36.653 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:21:36.653 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:21:36.653 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@309 -- # xtrace_disable 00:21:36.653 19:17:49 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # pci_devs=() 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # local -a pci_devs 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # pci_net_devs=() 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # pci_drivers=() 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # local -A pci_drivers 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # net_devs=() 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # local -ga net_devs 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # e810=() 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # local -ga e810 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # x722=() 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # local -ga x722 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # mlx=() 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # local -ga mlx 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:21:44.798 Found 0000:31:00.0 (0x8086 - 0x159b) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:21:44.798 Found 0000:31:00.1 (0x8086 - 0x159b) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@418 -- # [[ up == up ]] 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:21:44.798 Found net devices under 0000:31:00.0: cvl_0_0 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:21:44.798 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@418 -- # [[ up == up ]] 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:21:44.799 Found net devices under 0000:31:00.1: cvl_0_1 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # is_hw=yes 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:44.799 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:21:45.060 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:45.060 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.600 ms 00:21:45.060 00:21:45.060 --- 10.0.0.2 ping statistics --- 00:21:45.060 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:45.060 rtt min/avg/max/mdev = 0.600/0.600/0.600/0.000 ms 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:45.060 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:45.060 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.287 ms 00:21:45.060 00:21:45.060 --- 10.0.0.1 ping statistics --- 00:21:45.060 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:45.060 rtt min/avg/max/mdev = 0.287/0.287/0.287/0.000 ms 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@450 -- # return 0 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@13 -- # nvmfappstart 00:21:45.060 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@726 -- # xtrace_disable 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@509 -- # nvmfpid=3898293 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@510 -- # waitforlisten 3898293 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@835 -- # '[' -z 3898293 ']' 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:45.061 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:45.061 19:17:57 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:45.061 [2024-11-26 19:17:57.650754] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:45.061 [2024-11-26 19:17:57.650807] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:45.322 [2024-11-26 19:17:57.738709] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:45.322 [2024-11-26 19:17:57.775165] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:45.322 [2024-11-26 19:17:57.775203] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:45.322 [2024-11-26 19:17:57.775211] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:45.322 [2024-11-26 19:17:57.775218] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:45.322 [2024-11-26 19:17:57.775228] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:45.322 [2024-11-26 19:17:57.775785] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@868 -- # return 0 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@732 -- # xtrace_disable 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@19 -- # rpc_cmd nvmf_create_transport '-t tcp -o' --in-capsule-data-size 768 --control-msg-num 1 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:45.893 [2024-11-26 19:17:58.505181] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:45.893 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:46.154 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:46.154 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@21 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:46.155 Malloc0 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:46.155 [2024-11-26 19:17:58.556166] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@27 -- # perf_pid1=3898463 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x2 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@29 -- # perf_pid2=3898465 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x4 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@31 -- # perf_pid3=3898467 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@33 -- # wait 3898463 00:21:46.155 19:17:58 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x8 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:21:46.155 [2024-11-26 19:17:58.626550] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:21:46.155 [2024-11-26 19:17:58.656671] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:21:46.155 [2024-11-26 19:17:58.656914] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:21:47.540 Initializing NVMe Controllers 00:21:47.540 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:21:47.540 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 2 00:21:47.540 Initialization complete. Launching workers. 00:21:47.540 ======================================================== 00:21:47.540 Latency(us) 00:21:47.540 Device Information : IOPS MiB/s Average min max 00:21:47.540 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 2: 25.00 0.10 40880.85 40329.15 40950.54 00:21:47.540 ======================================================== 00:21:47.540 Total : 25.00 0.10 40880.85 40329.15 40950.54 00:21:47.540 00:21:47.540 Initializing NVMe Controllers 00:21:47.540 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:21:47.540 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 3 00:21:47.540 Initialization complete. Launching workers. 00:21:47.540 ======================================================== 00:21:47.540 Latency(us) 00:21:47.540 Device Information : IOPS MiB/s Average min max 00:21:47.540 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 3: 1526.00 5.96 655.49 311.83 845.54 00:21:47.540 ======================================================== 00:21:47.540 Total : 1526.00 5.96 655.49 311.83 845.54 00:21:47.540 00:21:47.540 Initializing NVMe Controllers 00:21:47.540 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:21:47.540 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 1 00:21:47.540 Initialization complete. Launching workers. 00:21:47.540 ======================================================== 00:21:47.540 Latency(us) 00:21:47.540 Device Information : IOPS MiB/s Average min max 00:21:47.540 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 1: 25.00 0.10 40923.94 40829.19 41439.12 00:21:47.540 ======================================================== 00:21:47.540 Total : 25.00 0.10 40923.94 40829.19 41439.12 00:21:47.540 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@34 -- # wait 3898465 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@35 -- # wait 3898467 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@38 -- # nvmftestfini 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@516 -- # nvmfcleanup 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@121 -- # sync 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@124 -- # set +e 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@125 -- # for i in {1..20} 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:21:47.540 rmmod nvme_tcp 00:21:47.540 rmmod nvme_fabrics 00:21:47.540 rmmod nvme_keyring 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@128 -- # set -e 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@129 -- # return 0 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@517 -- # '[' -n 3898293 ']' 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@518 -- # killprocess 3898293 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@954 -- # '[' -z 3898293 ']' 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@958 -- # kill -0 3898293 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@959 -- # uname 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3898293 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3898293' 00:21:47.540 killing process with pid 3898293 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@973 -- # kill 3898293 00:21:47.540 19:17:59 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@978 -- # wait 3898293 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@297 -- # iptr 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@791 -- # iptables-save 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@791 -- # iptables-restore 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@302 -- # remove_spdk_ns 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:47.540 19:18:00 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:49.585 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:21:49.585 00:21:49.585 real 0m13.159s 00:21:49.585 user 0m8.237s 00:21:49.585 sys 0m6.990s 00:21:49.585 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1130 -- # xtrace_disable 00:21:49.585 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:21:49.585 ************************************ 00:21:49.585 END TEST nvmf_control_msg_list 00:21:49.585 ************************************ 00:21:49.585 19:18:02 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@44 -- # run_test nvmf_wait_for_buf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:21:49.585 19:18:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:21:49.585 19:18:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:21:49.585 19:18:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:21:49.845 ************************************ 00:21:49.845 START TEST nvmf_wait_for_buf 00:21:49.845 ************************************ 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:21:49.845 * Looking for test storage... 00:21:49.845 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1693 -- # lcov --version 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # IFS=.-: 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # read -ra ver1 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # IFS=.-: 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # read -ra ver2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@338 -- # local 'op=<' 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@340 -- # ver1_l=2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@341 -- # ver2_l=1 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@344 -- # case "$op" in 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@345 -- # : 1 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v = 0 )) 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # decimal 1 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=1 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 1 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # ver1[v]=1 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # decimal 2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # ver2[v]=2 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # return 0 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:21:49.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:49.845 --rc genhtml_branch_coverage=1 00:21:49.845 --rc genhtml_function_coverage=1 00:21:49.845 --rc genhtml_legend=1 00:21:49.845 --rc geninfo_all_blocks=1 00:21:49.845 --rc geninfo_unexecuted_blocks=1 00:21:49.845 00:21:49.845 ' 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:21:49.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:49.845 --rc genhtml_branch_coverage=1 00:21:49.845 --rc genhtml_function_coverage=1 00:21:49.845 --rc genhtml_legend=1 00:21:49.845 --rc geninfo_all_blocks=1 00:21:49.845 --rc geninfo_unexecuted_blocks=1 00:21:49.845 00:21:49.845 ' 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:21:49.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:49.845 --rc genhtml_branch_coverage=1 00:21:49.845 --rc genhtml_function_coverage=1 00:21:49.845 --rc genhtml_legend=1 00:21:49.845 --rc geninfo_all_blocks=1 00:21:49.845 --rc geninfo_unexecuted_blocks=1 00:21:49.845 00:21:49.845 ' 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:21:49.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:49.845 --rc genhtml_branch_coverage=1 00:21:49.845 --rc genhtml_function_coverage=1 00:21:49.845 --rc genhtml_legend=1 00:21:49.845 --rc geninfo_all_blocks=1 00:21:49.845 --rc geninfo_unexecuted_blocks=1 00:21:49.845 00:21:49.845 ' 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # uname -s 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:49.845 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:49.846 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:49.846 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:49.846 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:49.846 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:49.846 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:49.846 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@15 -- # shopt -s extglob 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:50.106 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@5 -- # export PATH 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@51 -- # : 0 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:21:50.107 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@12 -- # nvmftestinit 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@476 -- # prepare_net_devs 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@438 -- # local -g is_hw=no 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # remove_spdk_ns 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@309 -- # xtrace_disable 00:21:50.107 19:18:02 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # pci_devs=() 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # local -a pci_devs 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # pci_drivers=() 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # net_devs=() 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # local -ga net_devs 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # e810=() 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # local -ga e810 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # x722=() 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # local -ga x722 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # mlx=() 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # local -ga mlx 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:21:58.250 Found 0000:31:00.0 (0x8086 - 0x159b) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:21:58.250 Found 0000:31:00.1 (0x8086 - 0x159b) 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:58.250 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:21:58.251 Found net devices under 0000:31:00.0: cvl_0_0 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:21:58.251 Found net devices under 0000:31:00.1: cvl_0_1 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # is_hw=yes 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:21:58.251 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:58.515 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:58.515 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:58.515 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:21:58.515 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:21:58.515 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:58.515 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.722 ms 00:21:58.515 00:21:58.515 --- 10.0.0.2 ping statistics --- 00:21:58.515 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:58.515 rtt min/avg/max/mdev = 0.722/0.722/0.722/0.000 ms 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:58.516 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:58.516 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.314 ms 00:21:58.516 00:21:58.516 --- 10.0.0.1 ping statistics --- 00:21:58.516 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:58.516 rtt min/avg/max/mdev = 0.314/0.314/0.314/0.000 ms 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@450 -- # return 0 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:21:58.516 19:18:10 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@13 -- # nvmfappstart --wait-for-rpc 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@726 -- # xtrace_disable 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@509 -- # nvmfpid=3904251 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@510 -- # waitforlisten 3904251 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@835 -- # '[' -z 3904251 ']' 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:58.517 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:58.517 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:58.517 [2024-11-26 19:18:11.082957] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:21:58.518 [2024-11-26 19:18:11.083023] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:58.788 [2024-11-26 19:18:11.176680] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:58.788 [2024-11-26 19:18:11.217507] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:58.788 [2024-11-26 19:18:11.217546] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:58.788 [2024-11-26 19:18:11.217554] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:58.788 [2024-11-26 19:18:11.217561] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:58.788 [2024-11-26 19:18:11.217566] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:58.788 [2024-11-26 19:18:11.218167] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@868 -- # return 0 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@732 -- # xtrace_disable 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@19 -- # rpc_cmd accel_set_options --small-cache-size 0 --large-cache-size 0 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@20 -- # rpc_cmd iobuf_set_options --small-pool-count 154 --small_bufsize=8192 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@21 -- # rpc_cmd framework_start_init 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.360 19:18:11 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@22 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.620 Malloc0 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@23 -- # rpc_cmd nvmf_create_transport '-t tcp -o' -u 8192 -n 24 -b 24 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.620 [2024-11-26 19:18:12.034366] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a -s SPDK00000000000001 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:21:59.620 [2024-11-26 19:18:12.070591] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:59.620 19:18:12 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 4 -o 131072 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:21:59.620 [2024-11-26 19:18:12.178950] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:22:00.999 Initializing NVMe Controllers 00:22:00.999 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:22:00.999 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 0 00:22:00.999 Initialization complete. Launching workers. 00:22:00.999 ======================================================== 00:22:00.999 Latency(us) 00:22:00.999 Device Information : IOPS MiB/s Average min max 00:22:00.999 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 0: 25.00 3.12 166321.11 47873.72 194545.04 00:22:00.999 ======================================================== 00:22:00.999 Total : 25.00 3.12 166321.11 47873.72 194545.04 00:22:00.999 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # rpc_cmd iobuf_get_stats 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # jq -r '.[] | select(.module == "nvmf_TCP") | .small_pool.retry' 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # retry_count=374 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@33 -- # [[ 374 -eq 0 ]] 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@38 -- # nvmftestfini 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@516 -- # nvmfcleanup 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@121 -- # sync 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@124 -- # set +e 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:00.999 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:00.999 rmmod nvme_tcp 00:22:01.261 rmmod nvme_fabrics 00:22:01.261 rmmod nvme_keyring 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@128 -- # set -e 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@129 -- # return 0 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@517 -- # '[' -n 3904251 ']' 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@518 -- # killprocess 3904251 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@954 -- # '[' -z 3904251 ']' 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@958 -- # kill -0 3904251 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@959 -- # uname 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3904251 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3904251' 00:22:01.261 killing process with pid 3904251 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@973 -- # kill 3904251 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@978 -- # wait 3904251 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@297 -- # iptr 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@791 -- # iptables-save 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@791 -- # iptables-restore 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:01.261 19:18:13 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:03.806 00:22:03.806 real 0m13.699s 00:22:03.806 user 0m5.354s 00:22:03.806 sys 0m6.911s 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:22:03.806 ************************************ 00:22:03.806 END TEST nvmf_wait_for_buf 00:22:03.806 ************************************ 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@47 -- # '[' 0 -eq 1 ']' 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@53 -- # [[ phy == phy ]] 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@54 -- # '[' tcp = tcp ']' 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@55 -- # gather_supported_nvmf_pci_devs 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@309 -- # xtrace_disable 00:22:03.806 19:18:15 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # pci_devs=() 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # net_devs=() 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # e810=() 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # local -ga e810 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # x722=() 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # local -ga x722 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # mlx=() 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # local -ga mlx 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:11.973 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:11.974 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:11.974 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:11.974 Found net devices under 0000:31:00.0: cvl_0_0 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:11.974 Found net devices under 0000:31:00.1: cvl_0_1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@56 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@57 -- # (( 2 > 0 )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@58 -- # run_test nvmf_perf_adq /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:22:11.974 ************************************ 00:22:11.974 START TEST nvmf_perf_adq 00:22:11.974 ************************************ 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:22:11.974 * Looking for test storage... 00:22:11.974 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1693 -- # lcov --version 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@333 -- # local ver1 ver1_l 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@334 -- # local ver2 ver2_l 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # IFS=.-: 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # read -ra ver1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # IFS=.-: 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # read -ra ver2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@338 -- # local 'op=<' 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@340 -- # ver1_l=2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@341 -- # ver2_l=1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@344 -- # case "$op" in 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@345 -- # : 1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v = 0 )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # decimal 1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # ver1[v]=1 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # decimal 2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # ver2[v]=2 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:22:11.974 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # return 0 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:22:11.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:11.975 --rc genhtml_branch_coverage=1 00:22:11.975 --rc genhtml_function_coverage=1 00:22:11.975 --rc genhtml_legend=1 00:22:11.975 --rc geninfo_all_blocks=1 00:22:11.975 --rc geninfo_unexecuted_blocks=1 00:22:11.975 00:22:11.975 ' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:22:11.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:11.975 --rc genhtml_branch_coverage=1 00:22:11.975 --rc genhtml_function_coverage=1 00:22:11.975 --rc genhtml_legend=1 00:22:11.975 --rc geninfo_all_blocks=1 00:22:11.975 --rc geninfo_unexecuted_blocks=1 00:22:11.975 00:22:11.975 ' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:22:11.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:11.975 --rc genhtml_branch_coverage=1 00:22:11.975 --rc genhtml_function_coverage=1 00:22:11.975 --rc genhtml_legend=1 00:22:11.975 --rc geninfo_all_blocks=1 00:22:11.975 --rc geninfo_unexecuted_blocks=1 00:22:11.975 00:22:11.975 ' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:22:11.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:11.975 --rc genhtml_branch_coverage=1 00:22:11.975 --rc genhtml_function_coverage=1 00:22:11.975 --rc genhtml_legend=1 00:22:11.975 --rc geninfo_all_blocks=1 00:22:11.975 --rc geninfo_unexecuted_blocks=1 00:22:11.975 00:22:11.975 ' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # uname -s 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@15 -- # shopt -s extglob 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@5 -- # export PATH 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@51 -- # : 0 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:22:11.975 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@55 -- # have_pci_nics=0 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@11 -- # gather_supported_nvmf_pci_devs 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:22:11.975 19:18:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:20.118 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:20.118 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:20.118 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:20.118 Found net devices under 0000:31:00.0: cvl_0_0 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:20.119 Found net devices under 0000:31:00.1: cvl_0_1 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@12 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@13 -- # (( 2 == 0 )) 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@18 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@68 -- # adq_reload_driver 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:22:20.119 19:18:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:22:21.500 19:18:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:22:24.042 19:18:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@76 -- # nvmftestinit 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # prepare_net_devs 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # local -g is_hw=no 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # remove_spdk_ns 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:29.322 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:29.323 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:29.323 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:29.323 Found net devices under 0000:31:00.0: cvl_0_0 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:29.323 Found net devices under 0000:31:00.1: cvl_0_1 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # is_hw=yes 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:29.323 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:29.324 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:29.324 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.659 ms 00:22:29.324 00:22:29.324 --- 10.0.0.2 ping statistics --- 00:22:29.324 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:29.324 rtt min/avg/max/mdev = 0.659/0.659/0.659/0.000 ms 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:29.324 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:29.324 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.285 ms 00:22:29.324 00:22:29.324 --- 10.0.0.1 ping statistics --- 00:22:29.324 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:29.324 rtt min/avg/max/mdev = 0.285/0.285/0.285/0.000 ms 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@450 -- # return 0 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@77 -- # nvmfappstart -m 0xF --wait-for-rpc 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@726 -- # xtrace_disable 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # nvmfpid=3915714 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # waitforlisten 3915714 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # '[' -z 3915714 ']' 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # local max_retries=100 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:29.324 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@844 -- # xtrace_disable 00:22:29.324 19:18:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.324 [2024-11-26 19:18:41.519614] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:22:29.324 [2024-11-26 19:18:41.519666] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:29.324 [2024-11-26 19:18:41.606567] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:29.324 [2024-11-26 19:18:41.643571] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:29.324 [2024-11-26 19:18:41.643604] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:29.324 [2024-11-26 19:18:41.643612] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:29.324 [2024-11-26 19:18:41.643618] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:29.324 [2024-11-26 19:18:41.643624] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:29.324 [2024-11-26 19:18:41.645149] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:29.324 [2024-11-26 19:18:41.645265] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:29.324 [2024-11-26 19:18:41.645419] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:29.324 [2024-11-26 19:18:41.645419] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:29.892 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:22:29.892 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@868 -- # return 0 00:22:29.892 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:22:29.892 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@732 -- # xtrace_disable 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@78 -- # adq_configure_nvmf_target 0 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 0 --enable-zerocopy-send-server -i posix 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 0 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:29.893 [2024-11-26 19:18:42.488372] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:29.893 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:30.152 Malloc1 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:30.152 [2024-11-26 19:18:42.561428] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@82 -- # perfpid=3915786 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@83 -- # sleep 2 00:22:30.152 19:18:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # rpc_cmd nvmf_get_stats 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # nvmf_stats='{ 00:22:32.063 "tick_rate": 2400000000, 00:22:32.063 "poll_groups": [ 00:22:32.063 { 00:22:32.063 "name": "nvmf_tgt_poll_group_000", 00:22:32.063 "admin_qpairs": 1, 00:22:32.063 "io_qpairs": 1, 00:22:32.063 "current_admin_qpairs": 1, 00:22:32.063 "current_io_qpairs": 1, 00:22:32.063 "pending_bdev_io": 0, 00:22:32.063 "completed_nvme_io": 19773, 00:22:32.063 "transports": [ 00:22:32.063 { 00:22:32.063 "trtype": "TCP" 00:22:32.063 } 00:22:32.063 ] 00:22:32.063 }, 00:22:32.063 { 00:22:32.063 "name": "nvmf_tgt_poll_group_001", 00:22:32.063 "admin_qpairs": 0, 00:22:32.063 "io_qpairs": 1, 00:22:32.063 "current_admin_qpairs": 0, 00:22:32.063 "current_io_qpairs": 1, 00:22:32.063 "pending_bdev_io": 0, 00:22:32.063 "completed_nvme_io": 27435, 00:22:32.063 "transports": [ 00:22:32.063 { 00:22:32.063 "trtype": "TCP" 00:22:32.063 } 00:22:32.063 ] 00:22:32.063 }, 00:22:32.063 { 00:22:32.063 "name": "nvmf_tgt_poll_group_002", 00:22:32.063 "admin_qpairs": 0, 00:22:32.063 "io_qpairs": 1, 00:22:32.063 "current_admin_qpairs": 0, 00:22:32.063 "current_io_qpairs": 1, 00:22:32.063 "pending_bdev_io": 0, 00:22:32.063 "completed_nvme_io": 21285, 00:22:32.063 "transports": [ 00:22:32.063 { 00:22:32.063 "trtype": "TCP" 00:22:32.063 } 00:22:32.063 ] 00:22:32.063 }, 00:22:32.063 { 00:22:32.063 "name": "nvmf_tgt_poll_group_003", 00:22:32.063 "admin_qpairs": 0, 00:22:32.063 "io_qpairs": 1, 00:22:32.063 "current_admin_qpairs": 0, 00:22:32.063 "current_io_qpairs": 1, 00:22:32.063 "pending_bdev_io": 0, 00:22:32.063 "completed_nvme_io": 20150, 00:22:32.063 "transports": [ 00:22:32.063 { 00:22:32.063 "trtype": "TCP" 00:22:32.063 } 00:22:32.063 ] 00:22:32.063 } 00:22:32.063 ] 00:22:32.063 }' 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 1) | length' 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # wc -l 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # count=4 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@87 -- # [[ 4 -ne 4 ]] 00:22:32.063 19:18:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@91 -- # wait 3915786 00:22:40.196 Initializing NVMe Controllers 00:22:40.196 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:22:40.196 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:22:40.196 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:22:40.196 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:22:40.196 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:22:40.196 Initialization complete. Launching workers. 00:22:40.196 ======================================================== 00:22:40.196 Latency(us) 00:22:40.196 Device Information : IOPS MiB/s Average min max 00:22:40.196 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 11319.81 44.22 5655.71 1740.76 8458.28 00:22:40.196 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 14306.76 55.89 4473.61 1165.10 9113.69 00:22:40.196 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 13994.27 54.67 4573.98 1245.30 10602.58 00:22:40.196 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 14063.57 54.94 4559.79 1352.18 45139.75 00:22:40.196 ======================================================== 00:22:40.196 Total : 53684.41 209.70 4771.61 1165.10 45139.75 00:22:40.196 00:22:40.196 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@92 -- # nvmftestfini 00:22:40.196 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # nvmfcleanup 00:22:40.196 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:22:40.196 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:40.196 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:22:40.196 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:40.196 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:40.196 rmmod nvme_tcp 00:22:40.196 rmmod nvme_fabrics 00:22:40.457 rmmod nvme_keyring 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@517 -- # '[' -n 3915714 ']' 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # killprocess 3915714 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # '[' -z 3915714 ']' 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@958 -- # kill -0 3915714 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # uname 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3915714 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3915714' 00:22:40.457 killing process with pid 3915714 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@973 -- # kill 3915714 00:22:40.457 19:18:52 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@978 -- # wait 3915714 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-save 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-restore 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:40.457 19:18:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:42.999 19:18:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:42.999 19:18:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@94 -- # adq_reload_driver 00:22:42.999 19:18:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:22:42.999 19:18:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:22:44.381 19:18:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:22:46.326 19:18:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@97 -- # nvmftestinit 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # prepare_net_devs 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # local -g is_hw=no 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # remove_spdk_ns 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:51.618 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:51.618 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:51.618 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:51.619 Found net devices under 0000:31:00.0: cvl_0_0 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:51.619 Found net devices under 0000:31:00.1: cvl_0_1 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # is_hw=yes 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:51.619 19:19:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:51.619 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:51.619 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.620 ms 00:22:51.619 00:22:51.619 --- 10.0.0.2 ping statistics --- 00:22:51.619 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:51.619 rtt min/avg/max/mdev = 0.620/0.620/0.620/0.000 ms 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:51.619 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:51.619 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.286 ms 00:22:51.619 00:22:51.619 --- 10.0.0.1 ping statistics --- 00:22:51.619 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:51.619 rtt min/avg/max/mdev = 0.286/0.286/0.286/0.000 ms 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@450 -- # return 0 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@98 -- # adq_configure_driver 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@22 -- # ip netns exec cvl_0_0_ns_spdk ethtool --offload cvl_0_0 hw-tc-offload on 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@24 -- # ip netns exec cvl_0_0_ns_spdk ethtool --set-priv-flags cvl_0_0 channel-pkt-inspect-optimize off 00:22:51.619 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@26 -- # sysctl -w net.core.busy_poll=1 00:22:51.881 net.core.busy_poll = 1 00:22:51.881 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@27 -- # sysctl -w net.core.busy_read=1 00:22:51.881 net.core.busy_read = 1 00:22:51.881 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@29 -- # tc=/usr/sbin/tc 00:22:51.881 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@31 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 root mqprio num_tc 2 map 0 1 queues 2@0 2@2 hw 1 mode channel 00:22:51.881 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 ingress 00:22:51.881 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@35 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc filter add dev cvl_0_0 protocol ip parent ffff: prio 1 flower dst_ip 10.0.0.2/32 ip_proto tcp dst_port 4420 skip_sw hw_tc 1 00:22:51.881 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@38 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/nvmf/set_xps_rxqs cvl_0_0 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@99 -- # nvmfappstart -m 0xF --wait-for-rpc 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@726 -- # xtrace_disable 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # nvmfpid=3920531 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # waitforlisten 3920531 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # '[' -z 3920531 ']' 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # local max_retries=100 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:52.142 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@844 -- # xtrace_disable 00:22:52.142 19:19:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:52.142 [2024-11-26 19:19:04.574014] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:22:52.142 [2024-11-26 19:19:04.574075] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:52.142 [2024-11-26 19:19:04.661663] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:52.142 [2024-11-26 19:19:04.701847] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:52.142 [2024-11-26 19:19:04.701889] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:52.142 [2024-11-26 19:19:04.701897] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:52.142 [2024-11-26 19:19:04.701904] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:52.142 [2024-11-26 19:19:04.701910] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:52.142 [2024-11-26 19:19:04.703586] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:52.142 [2024-11-26 19:19:04.703702] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:52.142 [2024-11-26 19:19:04.703858] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:52.142 [2024-11-26 19:19:04.703859] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@868 -- # return 0 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@732 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@100 -- # adq_configure_nvmf_target 1 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 1 --enable-zerocopy-send-server -i posix 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 1 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 [2024-11-26 19:19:05.546909] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 Malloc1 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:53.083 [2024-11-26 19:19:05.619340] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@104 -- # perfpid=3920788 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@105 -- # sleep 2 00:22:53.083 19:19:05 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # rpc_cmd nvmf_get_stats 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # nvmf_stats='{ 00:22:55.632 "tick_rate": 2400000000, 00:22:55.632 "poll_groups": [ 00:22:55.632 { 00:22:55.632 "name": "nvmf_tgt_poll_group_000", 00:22:55.632 "admin_qpairs": 1, 00:22:55.632 "io_qpairs": 2, 00:22:55.632 "current_admin_qpairs": 1, 00:22:55.632 "current_io_qpairs": 2, 00:22:55.632 "pending_bdev_io": 0, 00:22:55.632 "completed_nvme_io": 27903, 00:22:55.632 "transports": [ 00:22:55.632 { 00:22:55.632 "trtype": "TCP" 00:22:55.632 } 00:22:55.632 ] 00:22:55.632 }, 00:22:55.632 { 00:22:55.632 "name": "nvmf_tgt_poll_group_001", 00:22:55.632 "admin_qpairs": 0, 00:22:55.632 "io_qpairs": 2, 00:22:55.632 "current_admin_qpairs": 0, 00:22:55.632 "current_io_qpairs": 2, 00:22:55.632 "pending_bdev_io": 0, 00:22:55.632 "completed_nvme_io": 38546, 00:22:55.632 "transports": [ 00:22:55.632 { 00:22:55.632 "trtype": "TCP" 00:22:55.632 } 00:22:55.632 ] 00:22:55.632 }, 00:22:55.632 { 00:22:55.632 "name": "nvmf_tgt_poll_group_002", 00:22:55.632 "admin_qpairs": 0, 00:22:55.632 "io_qpairs": 0, 00:22:55.632 "current_admin_qpairs": 0, 00:22:55.632 "current_io_qpairs": 0, 00:22:55.632 "pending_bdev_io": 0, 00:22:55.632 "completed_nvme_io": 0, 00:22:55.632 "transports": [ 00:22:55.632 { 00:22:55.632 "trtype": "TCP" 00:22:55.632 } 00:22:55.632 ] 00:22:55.632 }, 00:22:55.632 { 00:22:55.632 "name": "nvmf_tgt_poll_group_003", 00:22:55.632 "admin_qpairs": 0, 00:22:55.632 "io_qpairs": 0, 00:22:55.632 "current_admin_qpairs": 0, 00:22:55.632 "current_io_qpairs": 0, 00:22:55.632 "pending_bdev_io": 0, 00:22:55.632 "completed_nvme_io": 0, 00:22:55.632 "transports": [ 00:22:55.632 { 00:22:55.632 "trtype": "TCP" 00:22:55.632 } 00:22:55.632 ] 00:22:55.632 } 00:22:55.632 ] 00:22:55.632 }' 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 0) | length' 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # wc -l 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # count=2 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@109 -- # [[ 2 -lt 2 ]] 00:22:55.632 19:19:07 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@114 -- # wait 3920788 00:23:03.766 Initializing NVMe Controllers 00:23:03.766 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:03.766 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:23:03.766 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:23:03.766 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:23:03.766 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:23:03.766 Initialization complete. Launching workers. 00:23:03.766 ======================================================== 00:23:03.766 Latency(us) 00:23:03.766 Device Information : IOPS MiB/s Average min max 00:23:03.766 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 11859.00 46.32 5398.58 987.83 51691.85 00:23:03.766 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 8606.50 33.62 7436.92 876.00 50136.64 00:23:03.766 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 9522.70 37.20 6719.98 1137.09 53887.61 00:23:03.766 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 9173.50 35.83 6983.34 1233.09 51132.38 00:23:03.766 ======================================================== 00:23:03.766 Total : 39161.68 152.98 6539.08 876.00 53887.61 00:23:03.766 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@115 -- # nvmftestfini 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # nvmfcleanup 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:03.766 rmmod nvme_tcp 00:23:03.766 rmmod nvme_fabrics 00:23:03.766 rmmod nvme_keyring 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@517 -- # '[' -n 3920531 ']' 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # killprocess 3920531 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # '[' -z 3920531 ']' 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@958 -- # kill -0 3920531 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # uname 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3920531 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3920531' 00:23:03.766 killing process with pid 3920531 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@973 -- # kill 3920531 00:23:03.766 19:19:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@978 -- # wait 3920531 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-restore 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-save 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:03.766 19:19:16 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@117 -- # trap - SIGINT SIGTERM EXIT 00:23:07.136 00:23:07.136 real 0m55.056s 00:23:07.136 user 2m49.857s 00:23:07.136 sys 0m12.311s 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1130 -- # xtrace_disable 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:23:07.136 ************************************ 00:23:07.136 END TEST nvmf_perf_adq 00:23:07.136 ************************************ 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@65 -- # run_test nvmf_shutdown /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:23:07.136 ************************************ 00:23:07.136 START TEST nvmf_shutdown 00:23:07.136 ************************************ 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:23:07.136 * Looking for test storage... 00:23:07.136 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1693 -- # lcov --version 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # IFS=.-: 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # read -ra ver1 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # IFS=.-: 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # read -ra ver2 00:23:07.136 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@338 -- # local 'op=<' 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@340 -- # ver1_l=2 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@341 -- # ver2_l=1 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@344 -- # case "$op" in 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@345 -- # : 1 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # decimal 1 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=1 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 1 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # ver1[v]=1 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # decimal 2 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=2 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 2 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # ver2[v]=2 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # return 0 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:23:07.137 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:07.137 --rc genhtml_branch_coverage=1 00:23:07.137 --rc genhtml_function_coverage=1 00:23:07.137 --rc genhtml_legend=1 00:23:07.137 --rc geninfo_all_blocks=1 00:23:07.137 --rc geninfo_unexecuted_blocks=1 00:23:07.137 00:23:07.137 ' 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:23:07.137 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:07.137 --rc genhtml_branch_coverage=1 00:23:07.137 --rc genhtml_function_coverage=1 00:23:07.137 --rc genhtml_legend=1 00:23:07.137 --rc geninfo_all_blocks=1 00:23:07.137 --rc geninfo_unexecuted_blocks=1 00:23:07.137 00:23:07.137 ' 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:23:07.137 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:07.137 --rc genhtml_branch_coverage=1 00:23:07.137 --rc genhtml_function_coverage=1 00:23:07.137 --rc genhtml_legend=1 00:23:07.137 --rc geninfo_all_blocks=1 00:23:07.137 --rc geninfo_unexecuted_blocks=1 00:23:07.137 00:23:07.137 ' 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:23:07.137 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:07.137 --rc genhtml_branch_coverage=1 00:23:07.137 --rc genhtml_function_coverage=1 00:23:07.137 --rc genhtml_legend=1 00:23:07.137 --rc geninfo_all_blocks=1 00:23:07.137 --rc geninfo_unexecuted_blocks=1 00:23:07.137 00:23:07.137 ' 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # uname -s 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@15 -- # shopt -s extglob 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:07.137 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@5 -- # export PATH 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@51 -- # : 0 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:07.138 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@12 -- # MALLOC_BDEV_SIZE=64 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@162 -- # run_test nvmf_shutdown_tc1 nvmf_shutdown_tc1 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:23:07.138 ************************************ 00:23:07.138 START TEST nvmf_shutdown_tc1 00:23:07.138 ************************************ 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc1 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@75 -- # starttarget 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@16 -- # nvmftestinit 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@476 -- # prepare_net_devs 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@309 -- # xtrace_disable 00:23:07.138 19:19:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # pci_devs=() 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # net_devs=() 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # e810=() 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # local -ga e810 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # x722=() 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # local -ga x722 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # mlx=() 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # local -ga mlx 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:15.364 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:15.364 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:15.364 Found net devices under 0000:31:00.0: cvl_0_0 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:15.364 Found net devices under 0000:31:00.1: cvl_0_1 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # is_hw=yes 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:15.364 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:15.365 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:15.365 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:15.365 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:15.365 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:15.365 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:15.365 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:15.365 19:19:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:15.625 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:15.625 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:15.625 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:15.625 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:15.625 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:15.625 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.546 ms 00:23:15.625 00:23:15.625 --- 10.0.0.2 ping statistics --- 00:23:15.625 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:15.625 rtt min/avg/max/mdev = 0.546/0.546/0.546/0.000 ms 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:15.626 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:15.626 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.294 ms 00:23:15.626 00:23:15.626 --- 10.0.0.1 ping statistics --- 00:23:15.626 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:15.626 rtt min/avg/max/mdev = 0.294/0.294/0.294/0.000 ms 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@450 -- # return 0 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@509 -- # nvmfpid=3927722 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@510 -- # waitforlisten 3927722 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # '[' -z 3927722 ']' 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # local max_retries=100 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:15.626 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@844 -- # xtrace_disable 00:23:15.626 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:15.626 [2024-11-26 19:19:28.176791] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:15.626 [2024-11-26 19:19:28.176872] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:15.886 [2024-11-26 19:19:28.289411] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:23:15.886 [2024-11-26 19:19:28.341969] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:15.886 [2024-11-26 19:19:28.342024] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:15.886 [2024-11-26 19:19:28.342033] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:15.886 [2024-11-26 19:19:28.342040] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:15.886 [2024-11-26 19:19:28.342047] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:15.886 [2024-11-26 19:19:28.344440] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:15.886 [2024-11-26 19:19:28.344610] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:23:15.886 [2024-11-26 19:19:28.344774] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:23:15.886 [2024-11-26 19:19:28.344776] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:16.456 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:23:16.456 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@868 -- # return 0 00:23:16.456 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:23:16.456 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:16.456 19:19:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:16.456 [2024-11-26 19:19:29.025718] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.456 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.457 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.457 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@36 -- # rpc_cmd 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:16.715 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:16.715 Malloc1 00:23:16.715 [2024-11-26 19:19:29.146566] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:16.715 Malloc2 00:23:16.715 Malloc3 00:23:16.715 Malloc4 00:23:16.715 Malloc5 00:23:16.715 Malloc6 00:23:16.975 Malloc7 00:23:16.975 Malloc8 00:23:16.975 Malloc9 00:23:16.975 Malloc10 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@79 -- # perfpid=3928102 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@80 -- # waitforlisten 3928102 /var/tmp/bdevperf.sock 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # '[' -z 3928102 ']' 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # local max_retries=100 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json /dev/fd/63 00:23:16.975 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@844 -- # xtrace_disable 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # config=() 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # local subsystem config 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:16.975 { 00:23:16.975 "params": { 00:23:16.975 "name": "Nvme$subsystem", 00:23:16.975 "trtype": "$TEST_TRANSPORT", 00:23:16.975 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:16.975 "adrfam": "ipv4", 00:23:16.975 "trsvcid": "$NVMF_PORT", 00:23:16.975 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:16.975 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:16.975 "hdgst": ${hdgst:-false}, 00:23:16.975 "ddgst": ${ddgst:-false} 00:23:16.975 }, 00:23:16.975 "method": "bdev_nvme_attach_controller" 00:23:16.975 } 00:23:16.975 EOF 00:23:16.975 )") 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:16.975 { 00:23:16.975 "params": { 00:23:16.975 "name": "Nvme$subsystem", 00:23:16.975 "trtype": "$TEST_TRANSPORT", 00:23:16.975 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:16.975 "adrfam": "ipv4", 00:23:16.975 "trsvcid": "$NVMF_PORT", 00:23:16.975 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:16.975 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:16.975 "hdgst": ${hdgst:-false}, 00:23:16.975 "ddgst": ${ddgst:-false} 00:23:16.975 }, 00:23:16.975 "method": "bdev_nvme_attach_controller" 00:23:16.975 } 00:23:16.975 EOF 00:23:16.975 )") 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:16.975 { 00:23:16.975 "params": { 00:23:16.975 "name": "Nvme$subsystem", 00:23:16.975 "trtype": "$TEST_TRANSPORT", 00:23:16.975 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:16.975 "adrfam": "ipv4", 00:23:16.975 "trsvcid": "$NVMF_PORT", 00:23:16.975 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:16.975 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:16.975 "hdgst": ${hdgst:-false}, 00:23:16.975 "ddgst": ${ddgst:-false} 00:23:16.975 }, 00:23:16.975 "method": "bdev_nvme_attach_controller" 00:23:16.975 } 00:23:16.975 EOF 00:23:16.975 )") 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:16.975 { 00:23:16.975 "params": { 00:23:16.975 "name": "Nvme$subsystem", 00:23:16.975 "trtype": "$TEST_TRANSPORT", 00:23:16.975 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:16.975 "adrfam": "ipv4", 00:23:16.975 "trsvcid": "$NVMF_PORT", 00:23:16.975 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:16.975 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:16.975 "hdgst": ${hdgst:-false}, 00:23:16.975 "ddgst": ${ddgst:-false} 00:23:16.975 }, 00:23:16.975 "method": "bdev_nvme_attach_controller" 00:23:16.975 } 00:23:16.975 EOF 00:23:16.975 )") 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:16.975 { 00:23:16.975 "params": { 00:23:16.975 "name": "Nvme$subsystem", 00:23:16.975 "trtype": "$TEST_TRANSPORT", 00:23:16.975 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:16.975 "adrfam": "ipv4", 00:23:16.975 "trsvcid": "$NVMF_PORT", 00:23:16.975 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:16.975 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:16.975 "hdgst": ${hdgst:-false}, 00:23:16.975 "ddgst": ${ddgst:-false} 00:23:16.975 }, 00:23:16.975 "method": "bdev_nvme_attach_controller" 00:23:16.975 } 00:23:16.975 EOF 00:23:16.975 )") 00:23:16.975 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:17.235 { 00:23:17.235 "params": { 00:23:17.235 "name": "Nvme$subsystem", 00:23:17.235 "trtype": "$TEST_TRANSPORT", 00:23:17.235 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:17.235 "adrfam": "ipv4", 00:23:17.235 "trsvcid": "$NVMF_PORT", 00:23:17.235 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:17.235 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:17.235 "hdgst": ${hdgst:-false}, 00:23:17.235 "ddgst": ${ddgst:-false} 00:23:17.235 }, 00:23:17.235 "method": "bdev_nvme_attach_controller" 00:23:17.235 } 00:23:17.235 EOF 00:23:17.235 )") 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:17.235 [2024-11-26 19:19:29.603525] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:17.235 [2024-11-26 19:19:29.603588] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:17.235 { 00:23:17.235 "params": { 00:23:17.235 "name": "Nvme$subsystem", 00:23:17.235 "trtype": "$TEST_TRANSPORT", 00:23:17.235 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:17.235 "adrfam": "ipv4", 00:23:17.235 "trsvcid": "$NVMF_PORT", 00:23:17.235 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:17.235 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:17.235 "hdgst": ${hdgst:-false}, 00:23:17.235 "ddgst": ${ddgst:-false} 00:23:17.235 }, 00:23:17.235 "method": "bdev_nvme_attach_controller" 00:23:17.235 } 00:23:17.235 EOF 00:23:17.235 )") 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:17.235 { 00:23:17.235 "params": { 00:23:17.235 "name": "Nvme$subsystem", 00:23:17.235 "trtype": "$TEST_TRANSPORT", 00:23:17.235 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:17.235 "adrfam": "ipv4", 00:23:17.235 "trsvcid": "$NVMF_PORT", 00:23:17.235 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:17.235 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:17.235 "hdgst": ${hdgst:-false}, 00:23:17.235 "ddgst": ${ddgst:-false} 00:23:17.235 }, 00:23:17.235 "method": "bdev_nvme_attach_controller" 00:23:17.235 } 00:23:17.235 EOF 00:23:17.235 )") 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:17.235 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:17.235 { 00:23:17.235 "params": { 00:23:17.235 "name": "Nvme$subsystem", 00:23:17.235 "trtype": "$TEST_TRANSPORT", 00:23:17.235 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:17.235 "adrfam": "ipv4", 00:23:17.235 "trsvcid": "$NVMF_PORT", 00:23:17.235 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:17.235 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:17.236 "hdgst": ${hdgst:-false}, 00:23:17.236 "ddgst": ${ddgst:-false} 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 } 00:23:17.236 EOF 00:23:17.236 )") 00:23:17.236 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:17.236 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:17.236 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:17.236 { 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme$subsystem", 00:23:17.236 "trtype": "$TEST_TRANSPORT", 00:23:17.236 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "$NVMF_PORT", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:17.236 "hdgst": ${hdgst:-false}, 00:23:17.236 "ddgst": ${ddgst:-false} 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 } 00:23:17.236 EOF 00:23:17.236 )") 00:23:17.236 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:17.236 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # jq . 00:23:17.236 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@585 -- # IFS=, 00:23:17.236 19:19:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme1", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme2", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme3", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme4", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme5", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme6", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme7", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme8", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme9", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 },{ 00:23:17.236 "params": { 00:23:17.236 "name": "Nvme10", 00:23:17.236 "trtype": "tcp", 00:23:17.236 "traddr": "10.0.0.2", 00:23:17.236 "adrfam": "ipv4", 00:23:17.236 "trsvcid": "4420", 00:23:17.236 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:23:17.236 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:23:17.236 "hdgst": false, 00:23:17.236 "ddgst": false 00:23:17.236 }, 00:23:17.236 "method": "bdev_nvme_attach_controller" 00:23:17.236 }' 00:23:17.236 [2024-11-26 19:19:29.691448] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:17.236 [2024-11-26 19:19:29.730126] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@868 -- # return 0 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@81 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@84 -- # kill -9 3928102 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@85 -- # rm -f /var/run/spdk_bdev1 00:23:18.617 19:19:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@88 -- # sleep 1 00:23:19.557 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 74: 3928102 Killed $rootdir/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json <(gen_nvmf_target_json "${num_subsystems[@]}") 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@89 -- # kill -0 3927722 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # config=() 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # local subsystem config 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.557 { 00:23:19.557 "params": { 00:23:19.557 "name": "Nvme$subsystem", 00:23:19.557 "trtype": "$TEST_TRANSPORT", 00:23:19.557 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.557 "adrfam": "ipv4", 00:23:19.557 "trsvcid": "$NVMF_PORT", 00:23:19.557 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.557 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.557 "hdgst": ${hdgst:-false}, 00:23:19.557 "ddgst": ${ddgst:-false} 00:23:19.557 }, 00:23:19.557 "method": "bdev_nvme_attach_controller" 00:23:19.557 } 00:23:19.557 EOF 00:23:19.557 )") 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.557 { 00:23:19.557 "params": { 00:23:19.557 "name": "Nvme$subsystem", 00:23:19.557 "trtype": "$TEST_TRANSPORT", 00:23:19.557 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.557 "adrfam": "ipv4", 00:23:19.557 "trsvcid": "$NVMF_PORT", 00:23:19.557 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.557 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.557 "hdgst": ${hdgst:-false}, 00:23:19.557 "ddgst": ${ddgst:-false} 00:23:19.557 }, 00:23:19.557 "method": "bdev_nvme_attach_controller" 00:23:19.557 } 00:23:19.557 EOF 00:23:19.557 )") 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.557 { 00:23:19.557 "params": { 00:23:19.557 "name": "Nvme$subsystem", 00:23:19.557 "trtype": "$TEST_TRANSPORT", 00:23:19.557 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.557 "adrfam": "ipv4", 00:23:19.557 "trsvcid": "$NVMF_PORT", 00:23:19.557 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.557 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.557 "hdgst": ${hdgst:-false}, 00:23:19.557 "ddgst": ${ddgst:-false} 00:23:19.557 }, 00:23:19.557 "method": "bdev_nvme_attach_controller" 00:23:19.557 } 00:23:19.557 EOF 00:23:19.557 )") 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.557 { 00:23:19.557 "params": { 00:23:19.557 "name": "Nvme$subsystem", 00:23:19.557 "trtype": "$TEST_TRANSPORT", 00:23:19.557 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.557 "adrfam": "ipv4", 00:23:19.557 "trsvcid": "$NVMF_PORT", 00:23:19.557 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.557 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.557 "hdgst": ${hdgst:-false}, 00:23:19.557 "ddgst": ${ddgst:-false} 00:23:19.557 }, 00:23:19.557 "method": "bdev_nvme_attach_controller" 00:23:19.557 } 00:23:19.557 EOF 00:23:19.557 )") 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.557 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.557 { 00:23:19.557 "params": { 00:23:19.557 "name": "Nvme$subsystem", 00:23:19.557 "trtype": "$TEST_TRANSPORT", 00:23:19.557 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.558 "adrfam": "ipv4", 00:23:19.558 "trsvcid": "$NVMF_PORT", 00:23:19.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.558 "hdgst": ${hdgst:-false}, 00:23:19.558 "ddgst": ${ddgst:-false} 00:23:19.558 }, 00:23:19.558 "method": "bdev_nvme_attach_controller" 00:23:19.558 } 00:23:19.558 EOF 00:23:19.558 )") 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.558 { 00:23:19.558 "params": { 00:23:19.558 "name": "Nvme$subsystem", 00:23:19.558 "trtype": "$TEST_TRANSPORT", 00:23:19.558 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.558 "adrfam": "ipv4", 00:23:19.558 "trsvcid": "$NVMF_PORT", 00:23:19.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.558 "hdgst": ${hdgst:-false}, 00:23:19.558 "ddgst": ${ddgst:-false} 00:23:19.558 }, 00:23:19.558 "method": "bdev_nvme_attach_controller" 00:23:19.558 } 00:23:19.558 EOF 00:23:19.558 )") 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.558 { 00:23:19.558 "params": { 00:23:19.558 "name": "Nvme$subsystem", 00:23:19.558 "trtype": "$TEST_TRANSPORT", 00:23:19.558 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.558 "adrfam": "ipv4", 00:23:19.558 "trsvcid": "$NVMF_PORT", 00:23:19.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.558 "hdgst": ${hdgst:-false}, 00:23:19.558 "ddgst": ${ddgst:-false} 00:23:19.558 }, 00:23:19.558 "method": "bdev_nvme_attach_controller" 00:23:19.558 } 00:23:19.558 EOF 00:23:19.558 )") 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.558 [2024-11-26 19:19:32.155978] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:19.558 [2024-11-26 19:19:32.156076] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3928689 ] 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.558 { 00:23:19.558 "params": { 00:23:19.558 "name": "Nvme$subsystem", 00:23:19.558 "trtype": "$TEST_TRANSPORT", 00:23:19.558 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.558 "adrfam": "ipv4", 00:23:19.558 "trsvcid": "$NVMF_PORT", 00:23:19.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.558 "hdgst": ${hdgst:-false}, 00:23:19.558 "ddgst": ${ddgst:-false} 00:23:19.558 }, 00:23:19.558 "method": "bdev_nvme_attach_controller" 00:23:19.558 } 00:23:19.558 EOF 00:23:19.558 )") 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.558 { 00:23:19.558 "params": { 00:23:19.558 "name": "Nvme$subsystem", 00:23:19.558 "trtype": "$TEST_TRANSPORT", 00:23:19.558 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.558 "adrfam": "ipv4", 00:23:19.558 "trsvcid": "$NVMF_PORT", 00:23:19.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.558 "hdgst": ${hdgst:-false}, 00:23:19.558 "ddgst": ${ddgst:-false} 00:23:19.558 }, 00:23:19.558 "method": "bdev_nvme_attach_controller" 00:23:19.558 } 00:23:19.558 EOF 00:23:19.558 )") 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:19.558 { 00:23:19.558 "params": { 00:23:19.558 "name": "Nvme$subsystem", 00:23:19.558 "trtype": "$TEST_TRANSPORT", 00:23:19.558 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:19.558 "adrfam": "ipv4", 00:23:19.558 "trsvcid": "$NVMF_PORT", 00:23:19.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:19.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:19.558 "hdgst": ${hdgst:-false}, 00:23:19.558 "ddgst": ${ddgst:-false} 00:23:19.558 }, 00:23:19.558 "method": "bdev_nvme_attach_controller" 00:23:19.558 } 00:23:19.558 EOF 00:23:19.558 )") 00:23:19.558 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:23:19.817 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # jq . 00:23:19.817 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@585 -- # IFS=, 00:23:19.817 19:19:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:23:19.817 "params": { 00:23:19.817 "name": "Nvme1", 00:23:19.817 "trtype": "tcp", 00:23:19.817 "traddr": "10.0.0.2", 00:23:19.817 "adrfam": "ipv4", 00:23:19.817 "trsvcid": "4420", 00:23:19.817 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:19.817 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:19.817 "hdgst": false, 00:23:19.817 "ddgst": false 00:23:19.817 }, 00:23:19.817 "method": "bdev_nvme_attach_controller" 00:23:19.817 },{ 00:23:19.817 "params": { 00:23:19.817 "name": "Nvme2", 00:23:19.817 "trtype": "tcp", 00:23:19.817 "traddr": "10.0.0.2", 00:23:19.817 "adrfam": "ipv4", 00:23:19.817 "trsvcid": "4420", 00:23:19.817 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme3", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme4", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme5", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme6", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme7", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme8", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme9", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 },{ 00:23:19.818 "params": { 00:23:19.818 "name": "Nvme10", 00:23:19.818 "trtype": "tcp", 00:23:19.818 "traddr": "10.0.0.2", 00:23:19.818 "adrfam": "ipv4", 00:23:19.818 "trsvcid": "4420", 00:23:19.818 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:23:19.818 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:23:19.818 "hdgst": false, 00:23:19.818 "ddgst": false 00:23:19.818 }, 00:23:19.818 "method": "bdev_nvme_attach_controller" 00:23:19.818 }' 00:23:19.818 [2024-11-26 19:19:32.240385] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:19.818 [2024-11-26 19:19:32.276783] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:21.195 Running I/O for 1 seconds... 00:23:22.394 1870.00 IOPS, 116.88 MiB/s 00:23:22.394 Latency(us) 00:23:22.394 [2024-11-26T18:19:35.019Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:22.394 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme1n1 : 1.12 229.03 14.31 0.00 0.00 276617.81 16711.68 263891.63 00:23:22.394 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme2n1 : 1.03 186.70 11.67 0.00 0.00 332890.45 17913.17 272629.76 00:23:22.394 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme3n1 : 1.13 226.94 14.18 0.00 0.00 269709.87 17585.49 244667.73 00:23:22.394 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme4n1 : 1.10 233.19 14.57 0.00 0.00 257385.39 36918.61 244667.73 00:23:22.394 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme5n1 : 1.10 232.75 14.55 0.00 0.00 253205.33 22391.47 251658.24 00:23:22.394 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme6n1 : 1.12 228.18 14.26 0.00 0.00 253914.67 17694.72 260396.37 00:23:22.394 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme7n1 : 1.16 280.37 17.52 0.00 0.00 203418.65 2894.51 221948.59 00:23:22.394 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme8n1 : 1.16 284.96 17.81 0.00 0.00 195736.96 1884.16 255153.49 00:23:22.394 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme9n1 : 1.20 267.37 16.71 0.00 0.00 206672.21 11796.48 248162.99 00:23:22.394 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:22.394 Verification LBA range: start 0x0 length 0x400 00:23:22.394 Nvme10n1 : 1.21 317.59 19.85 0.00 0.00 171018.60 10103.47 246415.36 00:23:22.394 [2024-11-26T18:19:35.019Z] =================================================================================================================== 00:23:22.394 [2024-11-26T18:19:35.019Z] Total : 2487.10 155.44 0.00 0.00 233828.45 1884.16 272629.76 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@95 -- # stoptarget 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@46 -- # nvmftestfini 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@516 -- # nvmfcleanup 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@121 -- # sync 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@124 -- # set +e 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:22.394 19:19:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:22.394 rmmod nvme_tcp 00:23:22.394 rmmod nvme_fabrics 00:23:22.394 rmmod nvme_keyring 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@128 -- # set -e 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@129 -- # return 0 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@517 -- # '[' -n 3927722 ']' 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@518 -- # killprocess 3927722 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@954 -- # '[' -z 3927722 ']' 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@958 -- # kill -0 3927722 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@959 -- # uname 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3927722 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3927722' 00:23:22.655 killing process with pid 3927722 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@973 -- # kill 3927722 00:23:22.655 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@978 -- # wait 3927722 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@297 -- # iptr 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@791 -- # iptables-save 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@791 -- # iptables-restore 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:22.916 19:19:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:24.831 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:24.831 00:23:24.831 real 0m17.910s 00:23:24.831 user 0m34.372s 00:23:24.831 sys 0m7.617s 00:23:24.831 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:23:24.831 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:23:24.831 ************************************ 00:23:24.831 END TEST nvmf_shutdown_tc1 00:23:24.831 ************************************ 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@163 -- # run_test nvmf_shutdown_tc2 nvmf_shutdown_tc2 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:23:25.092 ************************************ 00:23:25.092 START TEST nvmf_shutdown_tc2 00:23:25.092 ************************************ 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc2 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@100 -- # starttarget 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@16 -- # nvmftestinit 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@476 -- # prepare_net_devs 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@309 -- # xtrace_disable 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # pci_devs=() 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # net_devs=() 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # e810=() 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # local -ga e810 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # x722=() 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # local -ga x722 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # mlx=() 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # local -ga mlx 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:25.092 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:25.092 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:25.092 Found net devices under 0000:31:00.0: cvl_0_0 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:25.092 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:25.093 Found net devices under 0000:31:00.1: cvl_0_1 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # is_hw=yes 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:25.093 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:25.353 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:25.353 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:25.353 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:25.353 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:25.353 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:25.353 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.656 ms 00:23:25.353 00:23:25.353 --- 10.0.0.2 ping statistics --- 00:23:25.353 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:25.353 rtt min/avg/max/mdev = 0.656/0.656/0.656/0.000 ms 00:23:25.353 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:25.354 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:25.354 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.282 ms 00:23:25.354 00:23:25.354 --- 10.0.0.1 ping statistics --- 00:23:25.354 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:25.354 rtt min/avg/max/mdev = 0.282/0.282/0.282/0.000 ms 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@450 -- # return 0 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@509 -- # nvmfpid=3929911 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@510 -- # waitforlisten 3929911 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # '[' -z 3929911 ']' 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:25.354 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:25.354 19:19:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:23:25.354 [2024-11-26 19:19:37.920097] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:25.354 [2024-11-26 19:19:37.920166] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:25.614 [2024-11-26 19:19:38.021859] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:23:25.614 [2024-11-26 19:19:38.055728] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:25.615 [2024-11-26 19:19:38.055758] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:25.615 [2024-11-26 19:19:38.055764] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:25.615 [2024-11-26 19:19:38.055769] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:25.615 [2024-11-26 19:19:38.055773] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:25.615 [2024-11-26 19:19:38.057372] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:25.615 [2024-11-26 19:19:38.057504] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:23:25.615 [2024-11-26 19:19:38.057632] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:25.615 [2024-11-26 19:19:38.057633] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@868 -- # return 0 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:26.185 [2024-11-26 19:19:38.749143] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:26.185 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:26.186 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:23:26.447 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@36 -- # rpc_cmd 00:23:26.447 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:26.447 19:19:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:26.447 Malloc1 00:23:26.447 [2024-11-26 19:19:38.870666] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:26.447 Malloc2 00:23:26.447 Malloc3 00:23:26.447 Malloc4 00:23:26.447 Malloc5 00:23:26.447 Malloc6 00:23:26.711 Malloc7 00:23:26.711 Malloc8 00:23:26.711 Malloc9 00:23:26.711 Malloc10 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@104 -- # perfpid=3930240 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@105 -- # waitforlisten 3930240 /var/tmp/bdevperf.sock 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # '[' -z 3930240 ']' 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:26.711 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # config=() 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # local subsystem config 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.711 { 00:23:26.711 "params": { 00:23:26.711 "name": "Nvme$subsystem", 00:23:26.711 "trtype": "$TEST_TRANSPORT", 00:23:26.711 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.711 "adrfam": "ipv4", 00:23:26.711 "trsvcid": "$NVMF_PORT", 00:23:26.711 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.711 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.711 "hdgst": ${hdgst:-false}, 00:23:26.711 "ddgst": ${ddgst:-false} 00:23:26.711 }, 00:23:26.711 "method": "bdev_nvme_attach_controller" 00:23:26.711 } 00:23:26.711 EOF 00:23:26.711 )") 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.711 { 00:23:26.711 "params": { 00:23:26.711 "name": "Nvme$subsystem", 00:23:26.711 "trtype": "$TEST_TRANSPORT", 00:23:26.711 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.711 "adrfam": "ipv4", 00:23:26.711 "trsvcid": "$NVMF_PORT", 00:23:26.711 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.711 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.711 "hdgst": ${hdgst:-false}, 00:23:26.711 "ddgst": ${ddgst:-false} 00:23:26.711 }, 00:23:26.711 "method": "bdev_nvme_attach_controller" 00:23:26.711 } 00:23:26.711 EOF 00:23:26.711 )") 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.711 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.712 { 00:23:26.712 "params": { 00:23:26.712 "name": "Nvme$subsystem", 00:23:26.712 "trtype": "$TEST_TRANSPORT", 00:23:26.712 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.712 "adrfam": "ipv4", 00:23:26.712 "trsvcid": "$NVMF_PORT", 00:23:26.712 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.712 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.712 "hdgst": ${hdgst:-false}, 00:23:26.712 "ddgst": ${ddgst:-false} 00:23:26.712 }, 00:23:26.712 "method": "bdev_nvme_attach_controller" 00:23:26.712 } 00:23:26.712 EOF 00:23:26.712 )") 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.712 { 00:23:26.712 "params": { 00:23:26.712 "name": "Nvme$subsystem", 00:23:26.712 "trtype": "$TEST_TRANSPORT", 00:23:26.712 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.712 "adrfam": "ipv4", 00:23:26.712 "trsvcid": "$NVMF_PORT", 00:23:26.712 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.712 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.712 "hdgst": ${hdgst:-false}, 00:23:26.712 "ddgst": ${ddgst:-false} 00:23:26.712 }, 00:23:26.712 "method": "bdev_nvme_attach_controller" 00:23:26.712 } 00:23:26.712 EOF 00:23:26.712 )") 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.712 { 00:23:26.712 "params": { 00:23:26.712 "name": "Nvme$subsystem", 00:23:26.712 "trtype": "$TEST_TRANSPORT", 00:23:26.712 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.712 "adrfam": "ipv4", 00:23:26.712 "trsvcid": "$NVMF_PORT", 00:23:26.712 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.712 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.712 "hdgst": ${hdgst:-false}, 00:23:26.712 "ddgst": ${ddgst:-false} 00:23:26.712 }, 00:23:26.712 "method": "bdev_nvme_attach_controller" 00:23:26.712 } 00:23:26.712 EOF 00:23:26.712 )") 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.712 { 00:23:26.712 "params": { 00:23:26.712 "name": "Nvme$subsystem", 00:23:26.712 "trtype": "$TEST_TRANSPORT", 00:23:26.712 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.712 "adrfam": "ipv4", 00:23:26.712 "trsvcid": "$NVMF_PORT", 00:23:26.712 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.712 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.712 "hdgst": ${hdgst:-false}, 00:23:26.712 "ddgst": ${ddgst:-false} 00:23:26.712 }, 00:23:26.712 "method": "bdev_nvme_attach_controller" 00:23:26.712 } 00:23:26.712 EOF 00:23:26.712 )") 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.712 [2024-11-26 19:19:39.322332] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:26.712 [2024-11-26 19:19:39.322387] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3930240 ] 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.712 { 00:23:26.712 "params": { 00:23:26.712 "name": "Nvme$subsystem", 00:23:26.712 "trtype": "$TEST_TRANSPORT", 00:23:26.712 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.712 "adrfam": "ipv4", 00:23:26.712 "trsvcid": "$NVMF_PORT", 00:23:26.712 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.712 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.712 "hdgst": ${hdgst:-false}, 00:23:26.712 "ddgst": ${ddgst:-false} 00:23:26.712 }, 00:23:26.712 "method": "bdev_nvme_attach_controller" 00:23:26.712 } 00:23:26.712 EOF 00:23:26.712 )") 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.712 { 00:23:26.712 "params": { 00:23:26.712 "name": "Nvme$subsystem", 00:23:26.712 "trtype": "$TEST_TRANSPORT", 00:23:26.712 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.712 "adrfam": "ipv4", 00:23:26.712 "trsvcid": "$NVMF_PORT", 00:23:26.712 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.712 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.712 "hdgst": ${hdgst:-false}, 00:23:26.712 "ddgst": ${ddgst:-false} 00:23:26.712 }, 00:23:26.712 "method": "bdev_nvme_attach_controller" 00:23:26.712 } 00:23:26.712 EOF 00:23:26.712 )") 00:23:26.712 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.973 { 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme$subsystem", 00:23:26.973 "trtype": "$TEST_TRANSPORT", 00:23:26.973 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "$NVMF_PORT", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.973 "hdgst": ${hdgst:-false}, 00:23:26.973 "ddgst": ${ddgst:-false} 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 } 00:23:26.973 EOF 00:23:26.973 )") 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:26.973 { 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme$subsystem", 00:23:26.973 "trtype": "$TEST_TRANSPORT", 00:23:26.973 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "$NVMF_PORT", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:26.973 "hdgst": ${hdgst:-false}, 00:23:26.973 "ddgst": ${ddgst:-false} 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 } 00:23:26.973 EOF 00:23:26.973 )") 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@584 -- # jq . 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@585 -- # IFS=, 00:23:26.973 19:19:39 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme1", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme2", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme3", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme4", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme5", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme6", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme7", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme8", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme9", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 },{ 00:23:26.973 "params": { 00:23:26.973 "name": "Nvme10", 00:23:26.973 "trtype": "tcp", 00:23:26.973 "traddr": "10.0.0.2", 00:23:26.973 "adrfam": "ipv4", 00:23:26.973 "trsvcid": "4420", 00:23:26.973 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:23:26.973 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:23:26.973 "hdgst": false, 00:23:26.973 "ddgst": false 00:23:26.973 }, 00:23:26.973 "method": "bdev_nvme_attach_controller" 00:23:26.973 }' 00:23:26.973 [2024-11-26 19:19:39.401407] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:26.973 [2024-11-26 19:19:39.437939] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:28.887 Running I/O for 10 seconds... 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@868 -- # return 0 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@106 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@108 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@58 -- # local ret=1 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@59 -- # local i 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=131 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 131 -ge 100 ']' 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@65 -- # ret=0 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@66 -- # break 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@70 -- # return 0 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@111 -- # killprocess 3930240 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # '[' -z 3930240 ']' 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@958 -- # kill -0 3930240 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # uname 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3930240 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3930240' 00:23:29.459 killing process with pid 3930240 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@973 -- # kill 3930240 00:23:29.459 19:19:41 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@978 -- # wait 3930240 00:23:29.459 Received shutdown signal, test time was about 0.871193 seconds 00:23:29.459 00:23:29.459 Latency(us) 00:23:29.459 [2024-11-26T18:19:42.084Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:29.459 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.459 Verification LBA range: start 0x0 length 0x400 00:23:29.459 Nvme1n1 : 0.82 233.78 14.61 0.00 0.00 270072.60 20316.16 276125.01 00:23:29.459 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.459 Verification LBA range: start 0x0 length 0x400 00:23:29.459 Nvme2n1 : 0.85 226.42 14.15 0.00 0.00 272600.18 17148.59 258648.75 00:23:29.460 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme3n1 : 0.85 300.05 18.75 0.00 0.00 200238.93 18896.21 202724.69 00:23:29.460 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme4n1 : 0.87 225.21 14.08 0.00 0.00 248460.85 20534.61 246415.36 00:23:29.460 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme5n1 : 0.81 235.67 14.73 0.00 0.00 242584.46 35826.35 244667.73 00:23:29.460 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme6n1 : 0.84 228.93 14.31 0.00 0.00 244094.29 19114.67 256901.12 00:23:29.460 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme7n1 : 0.83 231.05 14.44 0.00 0.00 235279.93 33641.81 232434.35 00:23:29.460 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme8n1 : 0.84 227.80 14.24 0.00 0.00 232929.28 18677.76 249910.61 00:23:29.460 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme9n1 : 0.84 229.91 14.37 0.00 0.00 224023.04 20971.52 251658.24 00:23:29.460 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:29.460 Verification LBA range: start 0x0 length 0x400 00:23:29.460 Nvme10n1 : 0.85 225.55 14.10 0.00 0.00 223379.34 20425.39 274377.39 00:23:29.460 [2024-11-26T18:19:42.085Z] =================================================================================================================== 00:23:29.460 [2024-11-26T18:19:42.085Z] Total : 2364.37 147.77 0.00 0.00 238124.96 17148.59 276125.01 00:23:29.721 19:19:42 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@114 -- # sleep 1 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@115 -- # kill -0 3929911 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@117 -- # stoptarget 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@46 -- # nvmftestfini 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@516 -- # nvmfcleanup 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@121 -- # sync 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@124 -- # set +e 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:30.664 rmmod nvme_tcp 00:23:30.664 rmmod nvme_fabrics 00:23:30.664 rmmod nvme_keyring 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@128 -- # set -e 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@129 -- # return 0 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@517 -- # '[' -n 3929911 ']' 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@518 -- # killprocess 3929911 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # '[' -z 3929911 ']' 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@958 -- # kill -0 3929911 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # uname 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:23:30.664 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3929911 00:23:30.926 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:23:30.926 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:23:30.926 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3929911' 00:23:30.926 killing process with pid 3929911 00:23:30.926 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@973 -- # kill 3929911 00:23:30.926 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@978 -- # wait 3929911 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@297 -- # iptr 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@791 -- # iptables-save 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@791 -- # iptables-restore 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:31.187 19:19:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:33.100 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:33.100 00:23:33.100 real 0m8.149s 00:23:33.100 user 0m25.131s 00:23:33.100 sys 0m1.271s 00:23:33.100 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:23:33.100 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:23:33.100 ************************************ 00:23:33.100 END TEST nvmf_shutdown_tc2 00:23:33.100 ************************************ 00:23:33.100 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@164 -- # run_test nvmf_shutdown_tc3 nvmf_shutdown_tc3 00:23:33.100 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:23:33.100 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:23:33.100 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:23:33.362 ************************************ 00:23:33.362 START TEST nvmf_shutdown_tc3 00:23:33.362 ************************************ 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc3 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@122 -- # starttarget 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@16 -- # nvmftestinit 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@476 -- # prepare_net_devs 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@309 -- # xtrace_disable 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # pci_devs=() 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # net_devs=() 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # e810=() 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # local -ga e810 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # x722=() 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # local -ga x722 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # mlx=() 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # local -ga mlx 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:33.362 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:33.362 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:33.362 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:33.363 Found net devices under 0000:31:00.0: cvl_0_0 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:33.363 Found net devices under 0000:31:00.1: cvl_0_1 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # is_hw=yes 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:33.363 19:19:45 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:33.625 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:33.625 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.634 ms 00:23:33.625 00:23:33.625 --- 10.0.0.2 ping statistics --- 00:23:33.625 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:33.625 rtt min/avg/max/mdev = 0.634/0.634/0.634/0.000 ms 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:33.625 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:33.625 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.316 ms 00:23:33.625 00:23:33.625 --- 10.0.0.1 ping statistics --- 00:23:33.625 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:33.625 rtt min/avg/max/mdev = 0.316/0.316/0.316/0.000 ms 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@450 -- # return 0 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@509 -- # nvmfpid=3931683 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@510 -- # waitforlisten 3931683 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # '[' -z 3931683 ']' 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # local max_retries=100 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:33.625 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@844 -- # xtrace_disable 00:23:33.625 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:33.625 [2024-11-26 19:19:46.189593] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:33.625 [2024-11-26 19:19:46.189663] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:33.885 [2024-11-26 19:19:46.294938] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:23:33.885 [2024-11-26 19:19:46.333875] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:33.885 [2024-11-26 19:19:46.333916] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:33.885 [2024-11-26 19:19:46.333922] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:33.885 [2024-11-26 19:19:46.333927] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:33.885 [2024-11-26 19:19:46.333931] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:33.885 [2024-11-26 19:19:46.335722] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:33.885 [2024-11-26 19:19:46.335903] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:23:33.885 [2024-11-26 19:19:46.336084] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:33.885 [2024-11-26 19:19:46.336085] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:23:34.455 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:23:34.455 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@868 -- # return 0 00:23:34.455 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:23:34.455 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:34.455 19:19:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:34.455 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:34.455 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:23:34.455 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:34.455 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:34.456 [2024-11-26 19:19:47.033408] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.456 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@36 -- # rpc_cmd 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:34.718 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:34.718 Malloc1 00:23:34.718 [2024-11-26 19:19:47.153627] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:34.718 Malloc2 00:23:34.718 Malloc3 00:23:34.718 Malloc4 00:23:34.718 Malloc5 00:23:34.718 Malloc6 00:23:34.980 Malloc7 00:23:34.980 Malloc8 00:23:34.980 Malloc9 00:23:34.980 Malloc10 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@126 -- # perfpid=3931924 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@127 -- # waitforlisten 3931924 /var/tmp/bdevperf.sock 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # '[' -z 3931924 ']' 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # local max_retries=100 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:34.980 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@844 -- # xtrace_disable 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # config=() 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # local subsystem config 00:23:34.980 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:34.981 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:34.981 { 00:23:34.981 "params": { 00:23:34.981 "name": "Nvme$subsystem", 00:23:34.981 "trtype": "$TEST_TRANSPORT", 00:23:34.981 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:34.981 "adrfam": "ipv4", 00:23:34.981 "trsvcid": "$NVMF_PORT", 00:23:34.981 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:34.981 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:34.981 "hdgst": ${hdgst:-false}, 00:23:34.981 "ddgst": ${ddgst:-false} 00:23:34.981 }, 00:23:34.981 "method": "bdev_nvme_attach_controller" 00:23:34.981 } 00:23:34.981 EOF 00:23:34.981 )") 00:23:34.981 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:34.981 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:34.981 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:34.981 { 00:23:34.981 "params": { 00:23:34.981 "name": "Nvme$subsystem", 00:23:34.981 "trtype": "$TEST_TRANSPORT", 00:23:34.981 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:34.981 "adrfam": "ipv4", 00:23:34.981 "trsvcid": "$NVMF_PORT", 00:23:34.981 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:34.981 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:34.981 "hdgst": ${hdgst:-false}, 00:23:34.981 "ddgst": ${ddgst:-false} 00:23:34.981 }, 00:23:34.981 "method": "bdev_nvme_attach_controller" 00:23:34.981 } 00:23:34.981 EOF 00:23:34.981 )") 00:23:34.981 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:34.981 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:34.982 { 00:23:34.982 "params": { 00:23:34.982 "name": "Nvme$subsystem", 00:23:34.982 "trtype": "$TEST_TRANSPORT", 00:23:34.982 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:34.982 "adrfam": "ipv4", 00:23:34.982 "trsvcid": "$NVMF_PORT", 00:23:34.982 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:34.982 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:34.982 "hdgst": ${hdgst:-false}, 00:23:34.982 "ddgst": ${ddgst:-false} 00:23:34.982 }, 00:23:34.982 "method": "bdev_nvme_attach_controller" 00:23:34.982 } 00:23:34.982 EOF 00:23:34.982 )") 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:34.982 { 00:23:34.982 "params": { 00:23:34.982 "name": "Nvme$subsystem", 00:23:34.982 "trtype": "$TEST_TRANSPORT", 00:23:34.982 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:34.982 "adrfam": "ipv4", 00:23:34.982 "trsvcid": "$NVMF_PORT", 00:23:34.982 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:34.982 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:34.982 "hdgst": ${hdgst:-false}, 00:23:34.982 "ddgst": ${ddgst:-false} 00:23:34.982 }, 00:23:34.982 "method": "bdev_nvme_attach_controller" 00:23:34.982 } 00:23:34.982 EOF 00:23:34.982 )") 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:34.982 { 00:23:34.982 "params": { 00:23:34.982 "name": "Nvme$subsystem", 00:23:34.982 "trtype": "$TEST_TRANSPORT", 00:23:34.982 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:34.982 "adrfam": "ipv4", 00:23:34.982 "trsvcid": "$NVMF_PORT", 00:23:34.982 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:34.982 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:34.982 "hdgst": ${hdgst:-false}, 00:23:34.982 "ddgst": ${ddgst:-false} 00:23:34.982 }, 00:23:34.982 "method": "bdev_nvme_attach_controller" 00:23:34.982 } 00:23:34.982 EOF 00:23:34.982 )") 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:34.982 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:34.982 { 00:23:34.982 "params": { 00:23:34.982 "name": "Nvme$subsystem", 00:23:34.982 "trtype": "$TEST_TRANSPORT", 00:23:34.982 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:34.982 "adrfam": "ipv4", 00:23:34.982 "trsvcid": "$NVMF_PORT", 00:23:34.982 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:34.982 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:34.982 "hdgst": ${hdgst:-false}, 00:23:34.982 "ddgst": ${ddgst:-false} 00:23:34.982 }, 00:23:34.982 "method": "bdev_nvme_attach_controller" 00:23:34.982 } 00:23:34.982 EOF 00:23:34.982 )") 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:35.244 { 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme$subsystem", 00:23:35.244 "trtype": "$TEST_TRANSPORT", 00:23:35.244 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:35.244 "adrfam": "ipv4", 00:23:35.244 "trsvcid": "$NVMF_PORT", 00:23:35.244 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:35.244 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:35.244 "hdgst": ${hdgst:-false}, 00:23:35.244 "ddgst": ${ddgst:-false} 00:23:35.244 }, 00:23:35.244 "method": "bdev_nvme_attach_controller" 00:23:35.244 } 00:23:35.244 EOF 00:23:35.244 )") 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:35.244 { 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme$subsystem", 00:23:35.244 "trtype": "$TEST_TRANSPORT", 00:23:35.244 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:35.244 "adrfam": "ipv4", 00:23:35.244 "trsvcid": "$NVMF_PORT", 00:23:35.244 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:35.244 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:35.244 "hdgst": ${hdgst:-false}, 00:23:35.244 "ddgst": ${ddgst:-false} 00:23:35.244 }, 00:23:35.244 "method": "bdev_nvme_attach_controller" 00:23:35.244 } 00:23:35.244 EOF 00:23:35.244 )") 00:23:35.244 [2024-11-26 19:19:47.617816] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:35.244 [2024-11-26 19:19:47.617892] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3931924 ] 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:35.244 { 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme$subsystem", 00:23:35.244 "trtype": "$TEST_TRANSPORT", 00:23:35.244 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:35.244 "adrfam": "ipv4", 00:23:35.244 "trsvcid": "$NVMF_PORT", 00:23:35.244 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:35.244 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:35.244 "hdgst": ${hdgst:-false}, 00:23:35.244 "ddgst": ${ddgst:-false} 00:23:35.244 }, 00:23:35.244 "method": "bdev_nvme_attach_controller" 00:23:35.244 } 00:23:35.244 EOF 00:23:35.244 )") 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:23:35.244 { 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme$subsystem", 00:23:35.244 "trtype": "$TEST_TRANSPORT", 00:23:35.244 "traddr": "$NVMF_FIRST_TARGET_IP", 00:23:35.244 "adrfam": "ipv4", 00:23:35.244 "trsvcid": "$NVMF_PORT", 00:23:35.244 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:23:35.244 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:23:35.244 "hdgst": ${hdgst:-false}, 00:23:35.244 "ddgst": ${ddgst:-false} 00:23:35.244 }, 00:23:35.244 "method": "bdev_nvme_attach_controller" 00:23:35.244 } 00:23:35.244 EOF 00:23:35.244 )") 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@584 -- # jq . 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@585 -- # IFS=, 00:23:35.244 19:19:47 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme1", 00:23:35.244 "trtype": "tcp", 00:23:35.244 "traddr": "10.0.0.2", 00:23:35.244 "adrfam": "ipv4", 00:23:35.244 "trsvcid": "4420", 00:23:35.244 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:35.244 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:35.244 "hdgst": false, 00:23:35.244 "ddgst": false 00:23:35.244 }, 00:23:35.244 "method": "bdev_nvme_attach_controller" 00:23:35.244 },{ 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme2", 00:23:35.244 "trtype": "tcp", 00:23:35.244 "traddr": "10.0.0.2", 00:23:35.244 "adrfam": "ipv4", 00:23:35.244 "trsvcid": "4420", 00:23:35.244 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:23:35.244 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:23:35.244 "hdgst": false, 00:23:35.244 "ddgst": false 00:23:35.244 }, 00:23:35.244 "method": "bdev_nvme_attach_controller" 00:23:35.244 },{ 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme3", 00:23:35.244 "trtype": "tcp", 00:23:35.244 "traddr": "10.0.0.2", 00:23:35.244 "adrfam": "ipv4", 00:23:35.244 "trsvcid": "4420", 00:23:35.244 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:23:35.244 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:23:35.244 "hdgst": false, 00:23:35.244 "ddgst": false 00:23:35.244 }, 00:23:35.244 "method": "bdev_nvme_attach_controller" 00:23:35.244 },{ 00:23:35.244 "params": { 00:23:35.244 "name": "Nvme4", 00:23:35.244 "trtype": "tcp", 00:23:35.244 "traddr": "10.0.0.2", 00:23:35.244 "adrfam": "ipv4", 00:23:35.245 "trsvcid": "4420", 00:23:35.245 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:23:35.245 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:23:35.245 "hdgst": false, 00:23:35.245 "ddgst": false 00:23:35.245 }, 00:23:35.245 "method": "bdev_nvme_attach_controller" 00:23:35.245 },{ 00:23:35.245 "params": { 00:23:35.245 "name": "Nvme5", 00:23:35.245 "trtype": "tcp", 00:23:35.245 "traddr": "10.0.0.2", 00:23:35.245 "adrfam": "ipv4", 00:23:35.245 "trsvcid": "4420", 00:23:35.245 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:23:35.245 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:23:35.245 "hdgst": false, 00:23:35.245 "ddgst": false 00:23:35.245 }, 00:23:35.245 "method": "bdev_nvme_attach_controller" 00:23:35.245 },{ 00:23:35.245 "params": { 00:23:35.245 "name": "Nvme6", 00:23:35.245 "trtype": "tcp", 00:23:35.245 "traddr": "10.0.0.2", 00:23:35.245 "adrfam": "ipv4", 00:23:35.245 "trsvcid": "4420", 00:23:35.245 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:23:35.245 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:23:35.245 "hdgst": false, 00:23:35.245 "ddgst": false 00:23:35.245 }, 00:23:35.245 "method": "bdev_nvme_attach_controller" 00:23:35.245 },{ 00:23:35.245 "params": { 00:23:35.245 "name": "Nvme7", 00:23:35.245 "trtype": "tcp", 00:23:35.245 "traddr": "10.0.0.2", 00:23:35.245 "adrfam": "ipv4", 00:23:35.245 "trsvcid": "4420", 00:23:35.245 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:23:35.245 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:23:35.245 "hdgst": false, 00:23:35.245 "ddgst": false 00:23:35.245 }, 00:23:35.245 "method": "bdev_nvme_attach_controller" 00:23:35.245 },{ 00:23:35.245 "params": { 00:23:35.245 "name": "Nvme8", 00:23:35.245 "trtype": "tcp", 00:23:35.245 "traddr": "10.0.0.2", 00:23:35.245 "adrfam": "ipv4", 00:23:35.245 "trsvcid": "4420", 00:23:35.245 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:23:35.245 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:23:35.245 "hdgst": false, 00:23:35.245 "ddgst": false 00:23:35.245 }, 00:23:35.245 "method": "bdev_nvme_attach_controller" 00:23:35.245 },{ 00:23:35.245 "params": { 00:23:35.245 "name": "Nvme9", 00:23:35.245 "trtype": "tcp", 00:23:35.245 "traddr": "10.0.0.2", 00:23:35.245 "adrfam": "ipv4", 00:23:35.245 "trsvcid": "4420", 00:23:35.245 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:23:35.245 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:23:35.245 "hdgst": false, 00:23:35.245 "ddgst": false 00:23:35.245 }, 00:23:35.245 "method": "bdev_nvme_attach_controller" 00:23:35.245 },{ 00:23:35.245 "params": { 00:23:35.245 "name": "Nvme10", 00:23:35.245 "trtype": "tcp", 00:23:35.245 "traddr": "10.0.0.2", 00:23:35.245 "adrfam": "ipv4", 00:23:35.245 "trsvcid": "4420", 00:23:35.245 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:23:35.245 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:23:35.245 "hdgst": false, 00:23:35.245 "ddgst": false 00:23:35.245 }, 00:23:35.245 "method": "bdev_nvme_attach_controller" 00:23:35.245 }' 00:23:35.245 [2024-11-26 19:19:47.699071] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:35.245 [2024-11-26 19:19:47.735437] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:36.627 Running I/O for 10 seconds... 00:23:36.627 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:23:36.627 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@868 -- # return 0 00:23:36.627 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@128 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:23:36.627 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:36.627 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@131 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@133 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@58 -- # local ret=1 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # local i 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=3 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 3 -ge 100 ']' 00:23:36.888 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=67 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 67 -ge 100 ']' 00:23:37.149 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:23:37.410 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:23:37.410 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:23:37.410 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:23:37.410 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:23:37.410 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:37.410 19:19:49 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:37.410 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=131 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 131 -ge 100 ']' 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@65 -- # ret=0 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@66 -- # break 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@70 -- # return 0 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@136 -- # killprocess 3931683 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # '[' -z 3931683 ']' 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@958 -- # kill -0 3931683 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@959 -- # uname 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3931683 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3931683' 00:23:37.686 killing process with pid 3931683 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@973 -- # kill 3931683 00:23:37.686 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@978 -- # wait 3931683 00:23:37.686 [2024-11-26 19:19:50.110317] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21050 is same with the state(6) to be set 00:23:37.686 [2024-11-26 19:19:50.110387] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21050 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.110394] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21050 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.110399] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21050 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.110404] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21050 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111206] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111241] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111248] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111252] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111257] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111262] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111267] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111272] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111282] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111287] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111292] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111297] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111302] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111306] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111311] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111315] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111320] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111325] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111329] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111334] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111338] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111343] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111348] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111353] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111357] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111361] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111366] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111371] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111376] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111381] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111385] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111389] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111394] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111398] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111403] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111407] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111414] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111419] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111423] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111428] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111433] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111437] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111442] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111447] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111452] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111457] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111461] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111466] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111470] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111475] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111479] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111484] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111488] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111493] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111498] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111502] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111506] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111511] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111516] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111520] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111524] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.687 [2024-11-26 19:19:50.111529] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.111534] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa0a990 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112583] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112595] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112601] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112605] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112610] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112615] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112620] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112625] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112629] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112634] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112639] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112643] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112649] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112653] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112664] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112669] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112674] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112679] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112683] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112688] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112693] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112698] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112702] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112707] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112712] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112717] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112724] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112728] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112734] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112739] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112744] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112749] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112754] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112758] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112763] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112768] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112773] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112777] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112782] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112786] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112791] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112795] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112800] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112804] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112809] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112814] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112819] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112824] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112828] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112834] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112839] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112844] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112848] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112853] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112858] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112867] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112872] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112877] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112882] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112886] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112891] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.112895] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21540 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.113979] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21a10 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.114007] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21a10 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.114741] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.688 [2024-11-26 19:19:50.114765] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114771] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114776] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114781] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114786] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114790] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114796] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114800] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114805] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114810] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114815] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114820] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114824] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114829] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114834] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114839] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114847] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114852] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114858] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114866] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114871] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114876] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114880] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114885] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114890] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114895] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114900] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114905] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114910] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114915] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114919] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114924] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114928] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114933] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114938] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114943] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114948] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114953] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114958] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114962] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114967] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114972] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114976] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114983] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114988] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114993] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.114998] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115002] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115007] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115012] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115017] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115022] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115027] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115032] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115036] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115041] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115046] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115051] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115055] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115060] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115065] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115069] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa21f00 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115676] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115691] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.689 [2024-11-26 19:19:50.115697] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115702] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115707] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115712] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115718] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115723] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115728] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115736] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115741] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115746] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115751] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115755] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115760] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115766] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115771] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115776] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115781] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115786] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115791] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115795] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115800] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115805] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115810] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115815] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115820] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115825] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115830] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115834] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115839] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115844] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115849] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115853] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115858] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115866] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115875] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115881] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115885] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115890] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115894] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115899] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115904] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115908] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115912] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115917] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115922] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115926] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115931] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115935] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115940] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115945] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115950] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115955] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115959] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115964] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115969] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115974] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115978] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115983] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115987] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115992] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.115997] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22280 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117337] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117354] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117360] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117364] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117369] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117374] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117379] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.690 [2024-11-26 19:19:50.117384] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117388] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117393] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117398] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117402] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117407] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117412] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117416] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117421] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117426] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117430] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117435] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117440] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117444] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117449] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117453] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117458] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117462] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117467] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117471] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117479] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117484] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117489] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117494] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117499] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117503] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117508] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117512] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117517] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117521] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117526] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117531] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117536] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117540] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117545] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117550] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117554] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117559] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117564] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117568] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117573] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117578] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117582] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117587] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117592] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117596] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117601] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117606] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117611] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117616] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117620] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117624] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117629] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117633] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117638] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.117644] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22ad0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118284] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118298] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118303] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118309] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118314] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118318] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118323] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118328] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118333] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118338] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118344] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118348] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118353] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118358] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.691 [2024-11-26 19:19:50.118363] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118368] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118372] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118378] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118383] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118390] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118395] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118400] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118405] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118409] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118414] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118419] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118423] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118428] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118433] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118438] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118443] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118447] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118452] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118457] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118461] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118466] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118471] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118475] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118480] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118485] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118490] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118494] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118500] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118504] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118509] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118514] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118520] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118525] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118530] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118535] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118540] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118545] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118550] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118554] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118559] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118564] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118569] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118573] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118579] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118583] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118588] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118592] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.118597] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa22fa0 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119052] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119066] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119071] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119076] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119081] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119086] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119091] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119096] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.692 [2024-11-26 19:19:50.119101] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119106] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119110] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119117] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119122] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119127] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119131] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119136] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119140] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119146] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119150] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119155] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119160] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119165] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119169] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119174] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119179] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119183] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119187] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119192] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119197] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119202] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119207] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119212] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119216] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119221] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119225] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119230] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119236] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119240] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119246] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119251] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119255] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119260] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119265] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119270] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119275] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119279] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119284] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119289] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119293] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119298] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119302] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119307] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119311] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119316] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.119320] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.693 [2024-11-26 19:19:50.125924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.125959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.125976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.125984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.125994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.126012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.126028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.126051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.126068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.126084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.126102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.693 [2024-11-26 19:19:50.126118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.693 [2024-11-26 19:19:50.126126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126296] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.694 [2024-11-26 19:19:50.126666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.694 [2024-11-26 19:19:50.126675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.126985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.126993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.127009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.127026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.695 [2024-11-26 19:19:50.127042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127073] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:37.695 [2024-11-26 19:19:50.127299] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127327] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127342] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127364] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127379] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb34230 is same with the state(6) to be set 00:23:37.695 [2024-11-26 19:19:50.127419] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127437] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127452] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127467] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127482] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf5b910 is same with the state(6) to be set 00:23:37.695 [2024-11-26 19:19:50.127500] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.695 [2024-11-26 19:19:50.127508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.695 [2024-11-26 19:19:50.127516] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127531] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127546] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127561] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xfaaef0 is same with the state(6) to be set 00:23:37.696 [2024-11-26 19:19:50.127587] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127603] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127621] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127637] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127651] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf5c9b0 is same with the state(6) to be set 00:23:37.696 [2024-11-26 19:19:50.127676] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127693] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127708] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127724] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127738] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb27b10 is same with the state(6) to be set 00:23:37.696 [2024-11-26 19:19:50.127759] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127776] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127791] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127807] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127821] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb358b0 is same with the state(6) to be set 00:23:37.696 [2024-11-26 19:19:50.127844] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127860] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127886] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127901] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127916] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf65f10 is same with the state(6) to be set 00:23:37.696 [2024-11-26 19:19:50.127941] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127960] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127975] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.127990] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.127997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.128004] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb34430 is same with the state(6) to be set 00:23:37.696 [2024-11-26 19:19:50.128025] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.128034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.128042] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.128049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.128057] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.128064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.128072] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.696 [2024-11-26 19:19:50.128079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.696 [2024-11-26 19:19:50.128086] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf98400 is same with the state(6) to be set 00:23:37.696 [2024-11-26 19:19:50.128147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.696 [2024-11-26 19:19:50.128159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128341] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128362] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128369] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128376] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128382] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128388] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128394] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128400] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa23470 is same with the state(6) to be set 00:23:37.697 [2024-11-26 19:19:50.128404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.697 [2024-11-26 19:19:50.128574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.697 [2024-11-26 19:19:50.128583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128665] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.128716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.128723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.698 [2024-11-26 19:19:50.137768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.698 [2024-11-26 19:19:50.137778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.137786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.137795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.137803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.137812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.137821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.137831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.137838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.137847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.137855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.137881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.137890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.137899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.137907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.699 [2024-11-26 19:19:50.138597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.699 [2024-11-26 19:19:50.138605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.138983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.138992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.700 [2024-11-26 19:19:50.139217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.700 [2024-11-26 19:19:50.139224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.139234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.139242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.139251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.139258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.140811] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6, 1] resetting controller 00:23:37.701 [2024-11-26 19:19:50.140846] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf5b910 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.140897] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb34230 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.140935] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.701 [2024-11-26 19:19:50.140945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.140954] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.701 [2024-11-26 19:19:50.140962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.140970] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.701 [2024-11-26 19:19:50.140978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.140987] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:37.701 [2024-11-26 19:19:50.140994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.141001] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf91540 is same with the state(6) to be set 00:23:37.701 [2024-11-26 19:19:50.141024] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xfaaef0 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.141038] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf5c9b0 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.141055] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb27b10 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.141069] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb358b0 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.141082] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf65f10 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.141099] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb34430 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.141115] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf98400 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.143781] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:23:37.701 [2024-11-26 19:19:50.143807] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3, 1] resetting controller 00:23:37.701 [2024-11-26 19:19:50.144792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.701 [2024-11-26 19:19:50.144817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf5b910 with addr=10.0.0.2, port=4420 00:23:37.701 [2024-11-26 19:19:50.144826] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf5b910 is same with the state(6) to be set 00:23:37.701 [2024-11-26 19:19:50.145234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.701 [2024-11-26 19:19:50.145274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb27b10 with addr=10.0.0.2, port=4420 00:23:37.701 [2024-11-26 19:19:50.145285] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb27b10 is same with the state(6) to be set 00:23:37.701 [2024-11-26 19:19:50.145559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.701 [2024-11-26 19:19:50.145571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb358b0 with addr=10.0.0.2, port=4420 00:23:37.701 [2024-11-26 19:19:50.145579] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb358b0 is same with the state(6) to be set 00:23:37.701 [2024-11-26 19:19:50.145918] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:23:37.701 [2024-11-26 19:19:50.146233] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:23:37.701 [2024-11-26 19:19:50.146275] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:23:37.701 [2024-11-26 19:19:50.146323] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:23:37.701 [2024-11-26 19:19:50.146362] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:23:37.701 [2024-11-26 19:19:50.146441] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:23:37.701 [2024-11-26 19:19:50.146458] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf5b910 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.146471] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb27b10 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.146481] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb358b0 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.146603] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:23:37.701 [2024-11-26 19:19:50.146626] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Ctrlr is in error state 00:23:37.701 [2024-11-26 19:19:50.146635] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] controller reinitialization failed 00:23:37.701 [2024-11-26 19:19:50.146645] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] in failed state. 00:23:37.701 [2024-11-26 19:19:50.146654] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Resetting controller failed. 00:23:37.701 [2024-11-26 19:19:50.146663] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Ctrlr is in error state 00:23:37.701 [2024-11-26 19:19:50.146677] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] controller reinitialization failed 00:23:37.701 [2024-11-26 19:19:50.146685] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:23:37.701 [2024-11-26 19:19:50.146691] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller failed. 00:23:37.701 [2024-11-26 19:19:50.146699] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Ctrlr is in error state 00:23:37.701 [2024-11-26 19:19:50.146705] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] controller reinitialization failed 00:23:37.701 [2024-11-26 19:19:50.146712] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] in failed state. 00:23:37.701 [2024-11-26 19:19:50.146718] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Resetting controller failed. 00:23:37.701 [2024-11-26 19:19:50.150847] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf91540 (9): Bad file descriptor 00:23:37.701 [2024-11-26 19:19:50.151003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.151016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.151032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.151041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.151051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.151059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.151068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.151076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.151085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.151093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.701 [2024-11-26 19:19:50.151103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.701 [2024-11-26 19:19:50.151110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.702 [2024-11-26 19:19:50.151656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.702 [2024-11-26 19:19:50.151664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.151987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.151995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.152013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.152029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.152049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.152065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.152083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.152099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.703 [2024-11-26 19:19:50.152116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.703 [2024-11-26 19:19:50.152125] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd42c90 is same with the state(6) to be set 00:23:37.704 [2024-11-26 19:19:50.153406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153730] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.704 [2024-11-26 19:19:50.153941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.704 [2024-11-26 19:19:50.153950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.153957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.153967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.153975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.153986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.153993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.705 [2024-11-26 19:19:50.154413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.705 [2024-11-26 19:19:50.154421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.154431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.154438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.154447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.154455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.154465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.154472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.154481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.154489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.154498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.154506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.154515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.154522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.154531] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf3eb40 is same with the state(6) to be set 00:23:37.706 [2024-11-26 19:19:50.155800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.155983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.155991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.706 [2024-11-26 19:19:50.156192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.706 [2024-11-26 19:19:50.156199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.707 [2024-11-26 19:19:50.156735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.707 [2024-11-26 19:19:50.156742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.156919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.156927] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf3fe00 is same with the state(6) to be set 00:23:37.708 [2024-11-26 19:19:50.158197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.708 [2024-11-26 19:19:50.158530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.708 [2024-11-26 19:19:50.158537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.158988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.158997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.159005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.159014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.159021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.709 [2024-11-26 19:19:50.159031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.709 [2024-11-26 19:19:50.159039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:32768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:32896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:33024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:33152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.159312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.159320] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf42380 is same with the state(6) to be set 00:23:37.710 [2024-11-26 19:19:50.160594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160665] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.710 [2024-11-26 19:19:50.160804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.710 [2024-11-26 19:19:50.160811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.160985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.160995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.711 [2024-11-26 19:19:50.161337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.711 [2024-11-26 19:19:50.161344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161353] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.161701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.161710] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf43640 is same with the state(6) to be set 00:23:37.712 [2024-11-26 19:19:50.163005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.163034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.163055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.163075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.163096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.163114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.163131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.712 [2024-11-26 19:19:50.163148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.712 [2024-11-26 19:19:50.163156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.713 [2024-11-26 19:19:50.163650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.713 [2024-11-26 19:19:50.163658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.163985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.163992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.714 [2024-11-26 19:19:50.164130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.714 [2024-11-26 19:19:50.164138] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdc5e50 is same with the state(6) to be set 00:23:37.714 [2024-11-26 19:19:50.165393] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2, 1] resetting controller 00:23:37.714 [2024-11-26 19:19:50.165414] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4, 1] resetting controller 00:23:37.714 [2024-11-26 19:19:50.165427] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5, 1] resetting controller 00:23:37.714 [2024-11-26 19:19:50.165440] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] resetting controller 00:23:37.714 [2024-11-26 19:19:50.165528] bdev_nvme.c:3168:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode8, 1] Unable to perform failover, already in progress. 00:23:37.715 [2024-11-26 19:19:50.165546] bdev_nvme.c:3168:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] Unable to perform failover, already in progress. 00:23:37.715 [2024-11-26 19:19:50.165624] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode8, 1] resetting controller 00:23:37.715 [2024-11-26 19:19:50.165636] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] resetting controller 00:23:37.715 [2024-11-26 19:19:50.166109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.715 [2024-11-26 19:19:50.166150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb34430 with addr=10.0.0.2, port=4420 00:23:37.715 [2024-11-26 19:19:50.166164] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb34430 is same with the state(6) to be set 00:23:37.715 [2024-11-26 19:19:50.166551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.715 [2024-11-26 19:19:50.166563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb34230 with addr=10.0.0.2, port=4420 00:23:37.715 [2024-11-26 19:19:50.166571] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb34230 is same with the state(6) to be set 00:23:37.715 [2024-11-26 19:19:50.167092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.715 [2024-11-26 19:19:50.167130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf65f10 with addr=10.0.0.2, port=4420 00:23:37.715 [2024-11-26 19:19:50.167141] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf65f10 is same with the state(6) to be set 00:23:37.715 [2024-11-26 19:19:50.167498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.715 [2024-11-26 19:19:50.167509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf5c9b0 with addr=10.0.0.2, port=4420 00:23:37.715 [2024-11-26 19:19:50.167517] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf5c9b0 is same with the state(6) to be set 00:23:37.715 [2024-11-26 19:19:50.168927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.168942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.168959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.168967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.168983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.168990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.715 [2024-11-26 19:19:50.169282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.715 [2024-11-26 19:19:50.169291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169445] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.716 [2024-11-26 19:19:50.169828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.716 [2024-11-26 19:19:50.169840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.169983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.169991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.170000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.170008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.170017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.170025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.170034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:37.717 [2024-11-26 19:19:50.170041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:37.717 [2024-11-26 19:19:50.170049] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf44900 is same with the state(6) to be set 00:23:37.717 [2024-11-26 19:19:50.172215] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3, 1] resetting controller 00:23:37.717 [2024-11-26 19:19:50.172240] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:23:37.717 [2024-11-26 19:19:50.172250] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6, 1] resetting controller 00:23:37.717 task offset: 29440 on job bdev=Nvme6n1 fails 00:23:37.717 00:23:37.717 Latency(us) 00:23:37.717 [2024-11-26T18:19:50.342Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:37.717 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme1n1 ended in about 0.96 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme1n1 : 0.96 200.36 12.52 66.79 0.00 236850.77 16165.55 255153.49 00:23:37.717 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme2n1 ended in about 0.97 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme2n1 : 0.97 132.05 8.25 66.02 0.00 313044.20 20097.71 265639.25 00:23:37.717 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme3n1 ended in about 0.96 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme3n1 : 0.96 200.10 12.51 66.70 0.00 227444.27 17039.36 253405.87 00:23:37.717 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme4n1 ended in about 0.97 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme4n1 : 0.97 197.59 12.35 65.86 0.00 225619.95 9611.95 256901.12 00:23:37.717 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme5n1 ended in about 0.97 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme5n1 : 0.97 131.40 8.21 65.70 0.00 295254.76 20425.39 255153.49 00:23:37.717 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme6n1 ended in about 0.96 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme6n1 : 0.96 200.71 12.54 66.90 0.00 212190.51 13817.17 255153.49 00:23:37.717 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme7n1 ended in about 0.98 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme7n1 : 0.98 200.72 12.54 65.54 0.00 209003.55 15073.28 235929.60 00:23:37.717 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme8n1 ended in about 0.98 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme8n1 : 0.98 196.14 12.26 65.38 0.00 208045.87 18896.21 267386.88 00:23:37.717 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme9n1 ended in about 0.99 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme9n1 : 0.99 129.66 8.10 64.83 0.00 273771.52 41724.59 249910.61 00:23:37.717 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:23:37.717 Job: Nvme10n1 ended in about 0.98 seconds with error 00:23:37.717 Verification LBA range: start 0x0 length 0x400 00:23:37.717 Nvme10n1 : 0.98 130.44 8.15 65.22 0.00 265449.24 20862.29 270882.13 00:23:37.717 [2024-11-26T18:19:50.342Z] =================================================================================================================== 00:23:37.717 [2024-11-26T18:19:50.342Z] Total : 1719.15 107.45 658.94 0.00 242141.88 9611.95 270882.13 00:23:37.717 [2024-11-26 19:19:50.198376] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:23:37.717 [2024-11-26 19:19:50.198409] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode9, 1] resetting controller 00:23:37.717 [2024-11-26 19:19:50.198821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.717 [2024-11-26 19:19:50.198837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xfaaef0 with addr=10.0.0.2, port=4420 00:23:37.717 [2024-11-26 19:19:50.198853] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xfaaef0 is same with the state(6) to be set 00:23:37.717 [2024-11-26 19:19:50.199092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.717 [2024-11-26 19:19:50.199103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf98400 with addr=10.0.0.2, port=4420 00:23:37.717 [2024-11-26 19:19:50.199110] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf98400 is same with the state(6) to be set 00:23:37.717 [2024-11-26 19:19:50.199123] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb34430 (9): Bad file descriptor 00:23:37.717 [2024-11-26 19:19:50.199136] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb34230 (9): Bad file descriptor 00:23:37.717 [2024-11-26 19:19:50.199146] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf65f10 (9): Bad file descriptor 00:23:37.717 [2024-11-26 19:19:50.199155] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf5c9b0 (9): Bad file descriptor 00:23:37.717 [2024-11-26 19:19:50.199622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.717 [2024-11-26 19:19:50.199637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb358b0 with addr=10.0.0.2, port=4420 00:23:37.718 [2024-11-26 19:19:50.199644] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb358b0 is same with the state(6) to be set 00:23:37.718 [2024-11-26 19:19:50.199831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.718 [2024-11-26 19:19:50.199842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb27b10 with addr=10.0.0.2, port=4420 00:23:37.718 [2024-11-26 19:19:50.199849] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb27b10 is same with the state(6) to be set 00:23:37.718 [2024-11-26 19:19:50.200190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.718 [2024-11-26 19:19:50.200200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf5b910 with addr=10.0.0.2, port=4420 00:23:37.718 [2024-11-26 19:19:50.200208] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf5b910 is same with the state(6) to be set 00:23:37.718 [2024-11-26 19:19:50.200531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.718 [2024-11-26 19:19:50.200540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf91540 with addr=10.0.0.2, port=4420 00:23:37.718 [2024-11-26 19:19:50.200547] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf91540 is same with the state(6) to be set 00:23:37.718 [2024-11-26 19:19:50.200557] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xfaaef0 (9): Bad file descriptor 00:23:37.718 [2024-11-26 19:19:50.200567] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf98400 (9): Bad file descriptor 00:23:37.718 [2024-11-26 19:19:50.200576] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.200583] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.200593] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.200603] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.200611] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.200617] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.200624] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.200634] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.200641] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.200648] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.200655] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.200662] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.200669] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.200675] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.200682] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.200688] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.200729] bdev_nvme.c:3168:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] Unable to perform failover, already in progress. 00:23:37.718 [2024-11-26 19:19:50.200742] bdev_nvme.c:3168:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode8, 1] Unable to perform failover, already in progress. 00:23:37.718 [2024-11-26 19:19:50.201126] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb358b0 (9): Bad file descriptor 00:23:37.718 [2024-11-26 19:19:50.201140] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb27b10 (9): Bad file descriptor 00:23:37.718 [2024-11-26 19:19:50.201149] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf5b910 (9): Bad file descriptor 00:23:37.718 [2024-11-26 19:19:50.201158] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf91540 (9): Bad file descriptor 00:23:37.718 [2024-11-26 19:19:50.201167] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.201173] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.201180] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.201187] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.201195] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.201201] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.201208] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.201214] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.201464] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] resetting controller 00:23:37.718 [2024-11-26 19:19:50.201478] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5, 1] resetting controller 00:23:37.718 [2024-11-26 19:19:50.201487] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4, 1] resetting controller 00:23:37.718 [2024-11-26 19:19:50.201496] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2, 1] resetting controller 00:23:37.718 [2024-11-26 19:19:50.201533] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.201543] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.201551] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.201557] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.201564] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.201571] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.201578] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.201584] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.201591] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.201598] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.201605] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.201611] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.201618] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] Ctrlr is in error state 00:23:37.718 [2024-11-26 19:19:50.201624] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] controller reinitialization failed 00:23:37.718 [2024-11-26 19:19:50.201631] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] in failed state. 00:23:37.718 [2024-11-26 19:19:50.201638] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] Resetting controller failed. 00:23:37.718 [2024-11-26 19:19:50.201874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.718 [2024-11-26 19:19:50.201888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf5c9b0 with addr=10.0.0.2, port=4420 00:23:37.718 [2024-11-26 19:19:50.201896] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf5c9b0 is same with the state(6) to be set 00:23:37.718 [2024-11-26 19:19:50.202219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.719 [2024-11-26 19:19:50.202229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xf65f10 with addr=10.0.0.2, port=4420 00:23:37.719 [2024-11-26 19:19:50.202236] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf65f10 is same with the state(6) to be set 00:23:37.719 [2024-11-26 19:19:50.202560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.719 [2024-11-26 19:19:50.202570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb34230 with addr=10.0.0.2, port=4420 00:23:37.719 [2024-11-26 19:19:50.202577] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb34230 is same with the state(6) to be set 00:23:37.719 [2024-11-26 19:19:50.202958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:37.719 [2024-11-26 19:19:50.202967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb34430 with addr=10.0.0.2, port=4420 00:23:37.719 [2024-11-26 19:19:50.202975] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb34430 is same with the state(6) to be set 00:23:37.719 [2024-11-26 19:19:50.203005] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf5c9b0 (9): Bad file descriptor 00:23:37.719 [2024-11-26 19:19:50.203015] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xf65f10 (9): Bad file descriptor 00:23:37.719 [2024-11-26 19:19:50.203028] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb34230 (9): Bad file descriptor 00:23:37.719 [2024-11-26 19:19:50.203037] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb34430 (9): Bad file descriptor 00:23:37.719 [2024-11-26 19:19:50.203063] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Ctrlr is in error state 00:23:37.719 [2024-11-26 19:19:50.203070] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] controller reinitialization failed 00:23:37.719 [2024-11-26 19:19:50.203077] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] in failed state. 00:23:37.719 [2024-11-26 19:19:50.203084] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Resetting controller failed. 00:23:37.719 [2024-11-26 19:19:50.203091] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Ctrlr is in error state 00:23:37.719 [2024-11-26 19:19:50.203097] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] controller reinitialization failed 00:23:37.719 [2024-11-26 19:19:50.203104] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] in failed state. 00:23:37.719 [2024-11-26 19:19:50.203110] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Resetting controller failed. 00:23:37.719 [2024-11-26 19:19:50.203117] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Ctrlr is in error state 00:23:37.719 [2024-11-26 19:19:50.203123] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] controller reinitialization failed 00:23:37.719 [2024-11-26 19:19:50.203130] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] in failed state. 00:23:37.719 [2024-11-26 19:19:50.203136] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Resetting controller failed. 00:23:37.719 [2024-11-26 19:19:50.203143] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Ctrlr is in error state 00:23:37.719 [2024-11-26 19:19:50.203149] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] controller reinitialization failed 00:23:37.719 [2024-11-26 19:19:50.203156] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] in failed state. 00:23:37.719 [2024-11-26 19:19:50.203162] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Resetting controller failed. 00:23:37.980 19:19:50 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@137 -- # sleep 1 00:23:38.923 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@138 -- # NOT wait 3931924 00:23:38.923 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@652 -- # local es=0 00:23:38.923 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@654 -- # valid_exec_arg wait 3931924 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@640 -- # local arg=wait 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@644 -- # type -t wait 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@655 -- # wait 3931924 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@655 -- # es=255 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@664 -- # es=127 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@665 -- # case "$es" in 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@672 -- # es=1 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@140 -- # stoptarget 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@46 -- # nvmftestfini 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@516 -- # nvmfcleanup 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@121 -- # sync 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@124 -- # set +e 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:38.924 rmmod nvme_tcp 00:23:38.924 rmmod nvme_fabrics 00:23:38.924 rmmod nvme_keyring 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@128 -- # set -e 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@129 -- # return 0 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@517 -- # '[' -n 3931683 ']' 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@518 -- # killprocess 3931683 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # '[' -z 3931683 ']' 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@958 -- # kill -0 3931683 00:23:38.924 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (3931683) - No such process 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@981 -- # echo 'Process with pid 3931683 is not found' 00:23:38.924 Process with pid 3931683 is not found 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@297 -- # iptr 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@791 -- # iptables-save 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@791 -- # iptables-restore 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:38.924 19:19:51 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:41.495 00:23:41.495 real 0m7.820s 00:23:41.495 user 0m19.015s 00:23:41.495 sys 0m1.290s 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:23:41.495 ************************************ 00:23:41.495 END TEST nvmf_shutdown_tc3 00:23:41.495 ************************************ 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ e810 == \e\8\1\0 ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ tcp == \r\d\m\a ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@167 -- # run_test nvmf_shutdown_tc4 nvmf_shutdown_tc4 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:23:41.495 ************************************ 00:23:41.495 START TEST nvmf_shutdown_tc4 00:23:41.495 ************************************ 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc4 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@145 -- # starttarget 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@16 -- # nvmftestinit 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@476 -- # prepare_net_devs 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@309 -- # xtrace_disable 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # pci_devs=() 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # net_devs=() 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # e810=() 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # local -ga e810 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # x722=() 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # local -ga x722 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # mlx=() 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # local -ga mlx 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:41.495 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:41.495 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:41.496 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:41.496 Found net devices under 0000:31:00.0: cvl_0_0 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:41.496 Found net devices under 0000:31:00.1: cvl_0_1 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # is_hw=yes 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:41.496 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:41.496 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.656 ms 00:23:41.496 00:23:41.496 --- 10.0.0.2 ping statistics --- 00:23:41.496 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:41.496 rtt min/avg/max/mdev = 0.656/0.656/0.656/0.000 ms 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:41.496 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:41.496 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.283 ms 00:23:41.496 00:23:41.496 --- 10.0.0.1 ping statistics --- 00:23:41.496 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:41.496 rtt min/avg/max/mdev = 0.283/0.283/0.283/0.000 ms 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@450 -- # return 0 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:23:41.496 19:19:53 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@509 -- # nvmfpid=3933270 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@510 -- # waitforlisten 3933270 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@835 -- # '[' -z 3933270 ']' 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@840 -- # local max_retries=100 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:41.496 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@844 -- # xtrace_disable 00:23:41.496 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:41.496 [2024-11-26 19:19:54.088443] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:23:41.496 [2024-11-26 19:19:54.088490] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:41.757 [2024-11-26 19:19:54.180885] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:23:41.757 [2024-11-26 19:19:54.211235] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:41.757 [2024-11-26 19:19:54.211262] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:41.757 [2024-11-26 19:19:54.211267] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:41.757 [2024-11-26 19:19:54.211272] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:41.757 [2024-11-26 19:19:54.211276] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:41.757 [2024-11-26 19:19:54.212477] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:41.757 [2024-11-26 19:19:54.212636] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:23:41.757 [2024-11-26 19:19:54.212790] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:23:41.757 [2024-11-26 19:19:54.212791] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@868 -- # return 0 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:42.328 [2024-11-26 19:19:54.938630] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:42.328 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:23:42.589 19:19:54 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:23:42.589 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@36 -- # rpc_cmd 00:23:42.589 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:42.589 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:42.589 Malloc1 00:23:42.589 [2024-11-26 19:19:55.058705] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:42.589 Malloc2 00:23:42.589 Malloc3 00:23:42.589 Malloc4 00:23:42.589 Malloc5 00:23:42.849 Malloc6 00:23:42.849 Malloc7 00:23:42.849 Malloc8 00:23:42.849 Malloc9 00:23:42.849 Malloc10 00:23:42.849 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:42.849 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:23:42.849 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@732 -- # xtrace_disable 00:23:42.849 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:42.849 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@149 -- # perfpid=3933651 00:23:42.849 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@150 -- # sleep 5 00:23:42.849 19:19:55 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@148 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 45056 -O 4096 -w randwrite -t 20 -r 'trtype:tcp adrfam:IPV4 traddr:10.0.0.2 trsvcid:4420' -P 4 00:23:43.109 [2024-11-26 19:19:55.524931] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@152 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@155 -- # killprocess 3933270 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # '[' -z 3933270 ']' 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@958 -- # kill -0 3933270 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@959 -- # uname 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3933270 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3933270' 00:23:48.404 killing process with pid 3933270 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@973 -- # kill 3933270 00:23:48.404 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@978 -- # wait 3933270 00:23:48.404 [2024-11-26 19:20:00.540291] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540338] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540350] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540355] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540360] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540365] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540371] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540379] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae450 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540664] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae940 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540692] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae940 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540698] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeae940 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540816] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaee10 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540839] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaee10 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540844] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaee10 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540849] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaee10 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540854] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaee10 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.540859] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaee10 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541064] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeadf80 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541319] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead0b0 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541332] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead0b0 is same with the state(6) to be set 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 [2024-11-26 19:20:00.541519] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with the state(6) to be set 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 [2024-11-26 19:20:00.541534] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with the state(6) to be set 00:23:48.404 starting I/O failed: -6 00:23:48.404 [2024-11-26 19:20:00.541539] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with the state(6) to be set 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 [2024-11-26 19:20:00.541544] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541549] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541554] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with tWrite completed with error (sct=0, sc=8) 00:23:48.404 he state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541571] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541577] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xead5a0 is same with the state(6) to be set 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 [2024-11-26 19:20:00.541779] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeada90 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541792] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeada90 is same with the state(6) to be set 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 [2024-11-26 19:20:00.541797] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeada90 is same with the state(6) to be set 00:23:48.404 [2024-11-26 19:20:00.541802] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeada90 is same with the state(6) to be set 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 starting I/O failed: -6 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.404 Write completed with error (sct=0, sc=8) 00:23:48.405 [2024-11-26 19:20:00.542078] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeacbe0 is same with the state(6) to be set 00:23:48.405 [2024-11-26 19:20:00.542094] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeacbe0 is same with the state(6) to be set 00:23:48.405 [2024-11-26 19:20:00.542099] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeacbe0 is same with the state(6) to be set 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 [2024-11-26 19:20:00.542104] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeacbe0 is same with the state(6) to be set 00:23:48.405 starting I/O failed: -6 00:23:48.405 [2024-11-26 19:20:00.542113] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeacbe0 is same with the state(6) to be set 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 [2024-11-26 19:20:00.543269] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.405 Write completed with error (sct=0, sc=8) 00:23:48.405 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 [2024-11-26 19:20:00.545120] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.406 NVMe io qpair process completion error 00:23:48.406 [2024-11-26 19:20:00.548919] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf7b0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.548937] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf7b0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.548942] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf7b0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.548947] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf7b0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549234] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeafca0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549251] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeafca0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549526] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0170 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549543] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0170 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549548] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0170 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549553] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0170 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549558] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0170 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549746] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf2e0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549763] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf2e0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549769] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf2e0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.549775] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeaf2e0 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550183] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0b10 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550199] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0b10 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550204] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0b10 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550208] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0b10 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550213] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0b10 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550218] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0b10 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550424] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0be20 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550438] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0be20 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550676] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0c310 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550692] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0c310 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550697] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0c310 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550702] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0c310 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550707] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0c310 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550711] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe0c310 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550806] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0640 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550820] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0640 is same with the state(6) to be set 00:23:48.406 [2024-11-26 19:20:00.550825] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeb0640 is same with the state(6) to be set 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 [2024-11-26 19:20:00.552259] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 starting I/O failed: -6 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.406 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 [2024-11-26 19:20:00.553081] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.407 starting I/O failed: -6 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 [2024-11-26 19:20:00.554222] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.407 Write completed with error (sct=0, sc=8) 00:23:48.407 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 [2024-11-26 19:20:00.555671] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.408 NVMe io qpair process completion error 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 [2024-11-26 19:20:00.556690] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 [2024-11-26 19:20:00.557543] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.408 starting I/O failed: -6 00:23:48.408 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 [2024-11-26 19:20:00.558495] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 [2024-11-26 19:20:00.561249] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.409 NVMe io qpair process completion error 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 starting I/O failed: -6 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.409 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 [2024-11-26 19:20:00.562546] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 [2024-11-26 19:20:00.563449] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 [2024-11-26 19:20:00.564356] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.410 starting I/O failed: -6 00:23:48.410 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 [2024-11-26 19:20:00.566196] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.411 NVMe io qpair process completion error 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 [2024-11-26 19:20:00.567340] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.411 starting I/O failed: -6 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 [2024-11-26 19:20:00.568293] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.411 Write completed with error (sct=0, sc=8) 00:23:48.411 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 [2024-11-26 19:20:00.569217] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.412 Write completed with error (sct=0, sc=8) 00:23:48.412 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 [2024-11-26 19:20:00.571292] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.413 NVMe io qpair process completion error 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 [2024-11-26 19:20:00.572701] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 [2024-11-26 19:20:00.573529] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 [2024-11-26 19:20:00.574478] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.413 Write completed with error (sct=0, sc=8) 00:23:48.413 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 [2024-11-26 19:20:00.576472] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.414 NVMe io qpair process completion error 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 [2024-11-26 19:20:00.577524] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 starting I/O failed: -6 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.414 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 [2024-11-26 19:20:00.578430] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 [2024-11-26 19:20:00.579337] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.415 Write completed with error (sct=0, sc=8) 00:23:48.415 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 [2024-11-26 19:20:00.580991] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.416 NVMe io qpair process completion error 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 [2024-11-26 19:20:00.582130] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 [2024-11-26 19:20:00.583035] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 Write completed with error (sct=0, sc=8) 00:23:48.416 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 [2024-11-26 19:20:00.583949] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 [2024-11-26 19:20:00.587683] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.417 NVMe io qpair process completion error 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 [2024-11-26 19:20:00.588853] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.417 starting I/O failed: -6 00:23:48.417 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 [2024-11-26 19:20:00.589661] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 [2024-11-26 19:20:00.590601] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.418 Write completed with error (sct=0, sc=8) 00:23:48.418 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 [2024-11-26 19:20:00.592228] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:23:48.419 NVMe io qpair process completion error 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.419 starting I/O failed: -6 00:23:48.419 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 [2024-11-26 19:20:00.594566] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 Write completed with error (sct=0, sc=8) 00:23:48.420 starting I/O failed: -6 00:23:48.420 [2024-11-26 19:20:00.597438] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:23:48.420 NVMe io qpair process completion error 00:23:48.420 Initializing NVMe Controllers 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode4 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode3 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode7 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode10 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode8 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode9 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode6 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode2 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode5 00:23:48.420 Controller IO queue size 128, less than required. 00:23:48.420 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 with lcore 0 00:23:48.420 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 with lcore 0 00:23:48.420 Initialization complete. Launching workers. 00:23:48.420 ======================================================== 00:23:48.420 Latency(us) 00:23:48.420 Device Information : IOPS MiB/s Average min max 00:23:48.420 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 from core 0: 1867.05 80.22 68575.66 691.52 126651.58 00:23:48.420 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 from core 0: 1804.91 77.55 70970.57 614.04 151637.54 00:23:48.420 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 from core 0: 1857.09 79.80 69001.50 903.48 121832.34 00:23:48.420 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 from core 0: 1858.83 79.87 68961.37 865.60 131847.16 00:23:48.420 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1808.16 77.69 70227.11 660.73 120987.11 00:23:48.420 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 from core 0: 1859.26 79.89 68949.86 808.24 132848.98 00:23:48.421 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 from core 0: 1874.63 80.55 68397.19 594.46 135447.87 00:23:48.421 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 from core 0: 1869.65 80.34 67935.18 672.77 120050.66 00:23:48.421 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 from core 0: 1873.33 80.49 67820.13 670.10 121959.17 00:23:48.421 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 from core 0: 1816.17 78.04 69995.69 919.99 124258.59 00:23:48.421 ======================================================== 00:23:48.421 Total : 18489.09 794.45 69070.21 594.46 151637.54 00:23:48.421 00:23:48.421 [2024-11-26 19:20:00.601988] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f06c0 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602034] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f0390 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602064] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f1380 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602094] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f2540 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602121] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f2360 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602150] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f16b0 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602178] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f19e0 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602207] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f1050 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602238] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f0060 is same with the state(6) to be set 00:23:48.421 [2024-11-26 19:20:00.602266] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f09f0 is same with the state(6) to be set 00:23:48.421 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:23:48.421 19:20:00 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@156 -- # sleep 1 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@158 -- # NOT wait 3933651 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@652 -- # local es=0 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@654 -- # valid_exec_arg wait 3933651 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@640 -- # local arg=wait 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@644 -- # type -t wait 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@655 -- # wait 3933651 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@655 -- # es=1 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@159 -- # stoptarget 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@46 -- # nvmftestfini 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@516 -- # nvmfcleanup 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@121 -- # sync 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@124 -- # set +e 00:23:49.364 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:49.365 rmmod nvme_tcp 00:23:49.365 rmmod nvme_fabrics 00:23:49.365 rmmod nvme_keyring 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@128 -- # set -e 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@129 -- # return 0 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@517 -- # '[' -n 3933270 ']' 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@518 -- # killprocess 3933270 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # '[' -z 3933270 ']' 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@958 -- # kill -0 3933270 00:23:49.365 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (3933270) - No such process 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@981 -- # echo 'Process with pid 3933270 is not found' 00:23:49.365 Process with pid 3933270 is not found 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@297 -- # iptr 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@791 -- # iptables-save 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@791 -- # iptables-restore 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:49.365 19:20:01 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:51.907 19:20:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:51.907 00:23:51.907 real 0m10.332s 00:23:51.907 user 0m28.221s 00:23:51.907 sys 0m3.859s 00:23:51.907 19:20:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:23:51.907 19:20:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:23:51.907 ************************************ 00:23:51.907 END TEST nvmf_shutdown_tc4 00:23:51.907 ************************************ 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@170 -- # trap - SIGINT SIGTERM EXIT 00:23:51.907 00:23:51.907 real 0m44.753s 00:23:51.907 user 1m46.966s 00:23:51.907 sys 0m14.383s 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1130 -- # xtrace_disable 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:23:51.907 ************************************ 00:23:51.907 END TEST nvmf_shutdown 00:23:51.907 ************************************ 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@67 -- # run_test nvmf_nsid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nsid.sh --transport=tcp 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:23:51.907 ************************************ 00:23:51.907 START TEST nvmf_nsid 00:23:51.907 ************************************ 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nsid.sh --transport=tcp 00:23:51.907 * Looking for test storage... 00:23:51.907 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1693 -- # lcov --version 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@336 -- # IFS=.-: 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@336 -- # read -ra ver1 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@337 -- # IFS=.-: 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@337 -- # read -ra ver2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@338 -- # local 'op=<' 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@340 -- # ver1_l=2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@341 -- # ver2_l=1 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@344 -- # case "$op" in 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@345 -- # : 1 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@365 -- # decimal 1 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@353 -- # local d=1 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@355 -- # echo 1 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@365 -- # ver1[v]=1 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@366 -- # decimal 2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@353 -- # local d=2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@355 -- # echo 2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@366 -- # ver2[v]=2 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:51.907 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@368 -- # return 0 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:23:51.908 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:51.908 --rc genhtml_branch_coverage=1 00:23:51.908 --rc genhtml_function_coverage=1 00:23:51.908 --rc genhtml_legend=1 00:23:51.908 --rc geninfo_all_blocks=1 00:23:51.908 --rc geninfo_unexecuted_blocks=1 00:23:51.908 00:23:51.908 ' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:23:51.908 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:51.908 --rc genhtml_branch_coverage=1 00:23:51.908 --rc genhtml_function_coverage=1 00:23:51.908 --rc genhtml_legend=1 00:23:51.908 --rc geninfo_all_blocks=1 00:23:51.908 --rc geninfo_unexecuted_blocks=1 00:23:51.908 00:23:51.908 ' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:23:51.908 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:51.908 --rc genhtml_branch_coverage=1 00:23:51.908 --rc genhtml_function_coverage=1 00:23:51.908 --rc genhtml_legend=1 00:23:51.908 --rc geninfo_all_blocks=1 00:23:51.908 --rc geninfo_unexecuted_blocks=1 00:23:51.908 00:23:51.908 ' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:23:51.908 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:51.908 --rc genhtml_branch_coverage=1 00:23:51.908 --rc genhtml_function_coverage=1 00:23:51.908 --rc genhtml_legend=1 00:23:51.908 --rc geninfo_all_blocks=1 00:23:51.908 --rc geninfo_unexecuted_blocks=1 00:23:51.908 00:23:51.908 ' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@7 -- # uname -s 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@15 -- # shopt -s extglob 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@5 -- # export PATH 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@51 -- # : 0 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:51.908 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@11 -- # subnqn1=nqn.2024-10.io.spdk:cnode0 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@12 -- # subnqn2=nqn.2024-10.io.spdk:cnode1 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@13 -- # subnqn3=nqn.2024-10.io.spdk:cnode2 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@14 -- # tgt2sock=/var/tmp/tgt2.sock 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@15 -- # tgt2pid= 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@46 -- # nvmftestinit 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@476 -- # prepare_net_devs 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@438 -- # local -g is_hw=no 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@440 -- # remove_spdk_ns 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@309 -- # xtrace_disable 00:23:51.908 19:20:04 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@315 -- # pci_devs=() 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@319 -- # net_devs=() 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@320 -- # e810=() 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@320 -- # local -ga e810 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@321 -- # x722=() 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@321 -- # local -ga x722 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@322 -- # mlx=() 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@322 -- # local -ga mlx 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:24:00.046 Found 0000:31:00.0 (0x8086 - 0x159b) 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:00.046 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:24:00.047 Found 0000:31:00.1 (0x8086 - 0x159b) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:24:00.047 Found net devices under 0000:31:00.0: cvl_0_0 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:24:00.047 Found net devices under 0000:31:00.1: cvl_0_1 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@442 -- # is_hw=yes 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:00.047 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:00.308 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:00.308 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.645 ms 00:24:00.308 00:24:00.308 --- 10.0.0.2 ping statistics --- 00:24:00.308 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:00.308 rtt min/avg/max/mdev = 0.645/0.645/0.645/0.000 ms 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:00.308 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:00.308 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.289 ms 00:24:00.308 00:24:00.308 --- 10.0.0.1 ping statistics --- 00:24:00.308 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:00.308 rtt min/avg/max/mdev = 0.289/0.289/0.289/0.000 ms 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@450 -- # return 0 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@47 -- # nvmfappstart -m 1 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@509 -- # nvmfpid=3939610 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@510 -- # waitforlisten 3939610 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 1 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@835 -- # '[' -z 3939610 ']' 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:00.308 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:00.308 19:20:12 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:24:00.308 [2024-11-26 19:20:12.829096] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:00.308 [2024-11-26 19:20:12.829162] [ DPDK EAL parameters: nvmf -c 1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:00.308 [2024-11-26 19:20:12.918962] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:00.568 [2024-11-26 19:20:12.959740] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:00.568 [2024-11-26 19:20:12.959778] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:00.569 [2024-11-26 19:20:12.959786] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:00.569 [2024-11-26 19:20:12.959793] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:00.569 [2024-11-26 19:20:12.959799] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:00.569 [2024-11-26 19:20:12.960415] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@868 -- # return 0 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@732 -- # xtrace_disable 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@49 -- # trap cleanup SIGINT SIGTERM EXIT 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@52 -- # tgt2pid=3939716 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@54 -- # tgt1addr=10.0.0.2 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 2 -r /var/tmp/tgt2.sock 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@55 -- # get_main_ns_ip 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@769 -- # local ip 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@770 -- # ip_candidates=() 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@770 -- # local -A ip_candidates 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@55 -- # tgt2addr=10.0.0.1 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@56 -- # uuidgen 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@56 -- # ns1uuid=2535a682-feb1-4f78-af8c-e36cd507a1bb 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@57 -- # uuidgen 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@57 -- # ns2uuid=2561eec8-9716-4d96-bc80-baa4e0904f30 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@58 -- # uuidgen 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@58 -- # ns3uuid=ec6543bf-89bf-4fd2-8d69-fc35aff36503 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@63 -- # rpc_cmd 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:24:01.200 null0 00:24:01.200 null1 00:24:01.200 null2 00:24:01.200 [2024-11-26 19:20:13.727204] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:01.200 [2024-11-26 19:20:13.729522] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:01.200 [2024-11-26 19:20:13.729568] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3939716 ] 00:24:01.200 [2024-11-26 19:20:13.751428] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@79 -- # waitforlisten 3939716 /var/tmp/tgt2.sock 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@835 -- # '[' -z 3939716 ']' 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/tgt2.sock 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/tgt2.sock...' 00:24:01.200 Waiting for process to start up and listen on UNIX domain socket /var/tmp/tgt2.sock... 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:01.200 19:20:13 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:24:01.511 [2024-11-26 19:20:13.824674] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:01.511 [2024-11-26 19:20:13.860742] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:01.511 19:20:14 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:01.511 19:20:14 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@868 -- # return 0 00:24:01.511 19:20:14 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/tgt2.sock 00:24:01.787 [2024-11-26 19:20:14.354911] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:01.787 [2024-11-26 19:20:14.371037] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.1 port 4421 *** 00:24:01.787 nvme0n1 nvme0n2 00:24:01.787 nvme1n1 00:24:02.047 19:20:14 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@94 -- # nvme_connect 00:24:02.047 19:20:14 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@23 -- # local ctrlr 00:24:02.047 19:20:14 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@25 -- # nvme connect -t tcp -a 10.0.0.1 -s 4421 -n nqn.2024-10.io.spdk:cnode2 --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@28 -- # for ctrlr in /sys/class/nvme/nvme* 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@30 -- # [[ -e /sys/class/nvme/nvme0/subsysnqn ]] 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@30 -- # [[ nqn.2024-10.io.spdk:cnode2 == \n\q\n\.\2\0\2\4\-\1\0\.\i\o\.\s\p\d\k\:\c\n\o\d\e\2 ]] 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@31 -- # echo nvme0 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@32 -- # return 0 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@94 -- # ctrlr=nvme0 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@95 -- # waitforblk nvme0n1 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1239 -- # local i=0 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1241 -- # '[' 0 -lt 15 ']' 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1242 -- # i=1 00:24:03.431 19:20:15 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1243 -- # sleep 1 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n1 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1250 -- # return 0 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@96 -- # uuid2nguid 2535a682-feb1-4f78-af8c-e36cd507a1bb 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@787 -- # tr -d - 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@96 -- # nvme_get_nguid nvme0 1 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@40 -- # local ctrlr=nvme0 nsid=1 nguid 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nvme id-ns /dev/nvme0n1 -o json 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # jq -r .nguid 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nguid=2535a682feb14f78af8ce36cd507a1bb 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@43 -- # echo 2535A682FEB14F78AF8CE36CD507A1BB 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@96 -- # [[ 2535A682FEB14F78AF8CE36CD507A1BB == \2\5\3\5\A\6\8\2\F\E\B\1\4\F\7\8\A\F\8\C\E\3\6\C\D\5\0\7\A\1\B\B ]] 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@97 -- # waitforblk nvme0n2 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1239 -- # local i=0 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n2 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n2 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1250 -- # return 0 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@98 -- # uuid2nguid 2561eec8-9716-4d96-bc80-baa4e0904f30 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@787 -- # tr -d - 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@98 -- # nvme_get_nguid nvme0 2 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@40 -- # local ctrlr=nvme0 nsid=2 nguid 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nvme id-ns /dev/nvme0n2 -o json 00:24:04.373 19:20:16 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # jq -r .nguid 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nguid=2561eec897164d96bc80baa4e0904f30 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@43 -- # echo 2561EEC897164D96BC80BAA4E0904F30 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@98 -- # [[ 2561EEC897164D96BC80BAA4E0904F30 == \2\5\6\1\E\E\C\8\9\7\1\6\4\D\9\6\B\C\8\0\B\A\A\4\E\0\9\0\4\F\3\0 ]] 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@99 -- # waitforblk nvme0n3 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1239 -- # local i=0 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n3 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n3 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1250 -- # return 0 00:24:04.633 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@100 -- # uuid2nguid ec6543bf-89bf-4fd2-8d69-fc35aff36503 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@787 -- # tr -d - 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@100 -- # nvme_get_nguid nvme0 3 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@40 -- # local ctrlr=nvme0 nsid=3 nguid 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nvme id-ns /dev/nvme0n3 -o json 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # jq -r .nguid 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nguid=ec6543bf89bf4fd28d69fc35aff36503 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@43 -- # echo EC6543BF89BF4FD28D69FC35AFF36503 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@100 -- # [[ EC6543BF89BF4FD28D69FC35AFF36503 == \E\C\6\5\4\3\B\F\8\9\B\F\4\F\D\2\8\D\6\9\F\C\3\5\A\F\F\3\6\5\0\3 ]] 00:24:04.634 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@101 -- # nvme disconnect -d /dev/nvme0 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@103 -- # trap - SIGINT SIGTERM EXIT 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@104 -- # cleanup 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@18 -- # killprocess 3939716 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@954 -- # '[' -z 3939716 ']' 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@958 -- # kill -0 3939716 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # uname 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3939716 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3939716' 00:24:04.895 killing process with pid 3939716 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@973 -- # kill 3939716 00:24:04.895 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@978 -- # wait 3939716 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@19 -- # nvmftestfini 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@516 -- # nvmfcleanup 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@121 -- # sync 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@124 -- # set +e 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:05.156 rmmod nvme_tcp 00:24:05.156 rmmod nvme_fabrics 00:24:05.156 rmmod nvme_keyring 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@128 -- # set -e 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@129 -- # return 0 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@517 -- # '[' -n 3939610 ']' 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@518 -- # killprocess 3939610 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@954 -- # '[' -z 3939610 ']' 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@958 -- # kill -0 3939610 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # uname 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3939610 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3939610' 00:24:05.156 killing process with pid 3939610 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@973 -- # kill 3939610 00:24:05.156 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@978 -- # wait 3939610 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@297 -- # iptr 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@791 -- # iptables-save 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@791 -- # iptables-restore 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:05.417 19:20:17 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:07.961 19:20:19 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:07.961 00:24:07.961 real 0m15.870s 00:24:07.961 user 0m11.615s 00:24:07.961 sys 0m7.502s 00:24:07.961 19:20:19 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:07.961 19:20:19 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:24:07.961 ************************************ 00:24:07.961 END TEST nvmf_nsid 00:24:07.961 ************************************ 00:24:07.961 19:20:20 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:24:07.961 00:24:07.961 real 13m33.611s 00:24:07.961 user 27m46.337s 00:24:07.961 sys 4m11.445s 00:24:07.961 19:20:20 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:07.961 19:20:20 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:24:07.961 ************************************ 00:24:07.961 END TEST nvmf_target_extra 00:24:07.961 ************************************ 00:24:07.961 19:20:20 nvmf_tcp -- nvmf/nvmf.sh@16 -- # run_test nvmf_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:24:07.961 19:20:20 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:24:07.961 19:20:20 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:07.961 19:20:20 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:24:07.961 ************************************ 00:24:07.961 START TEST nvmf_host 00:24:07.961 ************************************ 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:24:07.961 * Looking for test storage... 00:24:07.961 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1693 -- # lcov --version 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # IFS=.-: 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # read -ra ver1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # IFS=.-: 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # read -ra ver2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@338 -- # local 'op=<' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@340 -- # ver1_l=2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@341 -- # ver2_l=1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@344 -- # case "$op" in 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@345 -- # : 1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # decimal 1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # ver1[v]=1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # decimal 2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # ver2[v]=2 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # return 0 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:07.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.961 --rc genhtml_branch_coverage=1 00:24:07.961 --rc genhtml_function_coverage=1 00:24:07.961 --rc genhtml_legend=1 00:24:07.961 --rc geninfo_all_blocks=1 00:24:07.961 --rc geninfo_unexecuted_blocks=1 00:24:07.961 00:24:07.961 ' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:07.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.961 --rc genhtml_branch_coverage=1 00:24:07.961 --rc genhtml_function_coverage=1 00:24:07.961 --rc genhtml_legend=1 00:24:07.961 --rc geninfo_all_blocks=1 00:24:07.961 --rc geninfo_unexecuted_blocks=1 00:24:07.961 00:24:07.961 ' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:07.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.961 --rc genhtml_branch_coverage=1 00:24:07.961 --rc genhtml_function_coverage=1 00:24:07.961 --rc genhtml_legend=1 00:24:07.961 --rc geninfo_all_blocks=1 00:24:07.961 --rc geninfo_unexecuted_blocks=1 00:24:07.961 00:24:07.961 ' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:07.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.961 --rc genhtml_branch_coverage=1 00:24:07.961 --rc genhtml_function_coverage=1 00:24:07.961 --rc genhtml_legend=1 00:24:07.961 --rc geninfo_all_blocks=1 00:24:07.961 --rc geninfo_unexecuted_blocks=1 00:24:07.961 00:24:07.961 ' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # uname -s 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@15 -- # shopt -s extglob 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:07.961 19:20:20 nvmf_tcp.nvmf_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- paths/export.sh@5 -- # export PATH 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@51 -- # : 0 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:07.962 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@13 -- # TEST_ARGS=("$@") 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@15 -- # [[ 0 -eq 0 ]] 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@16 -- # run_test nvmf_multicontroller /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:24:07.962 ************************************ 00:24:07.962 START TEST nvmf_multicontroller 00:24:07.962 ************************************ 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:24:07.962 * Looking for test storage... 00:24:07.962 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1693 -- # lcov --version 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # IFS=.-: 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # read -ra ver1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # IFS=.-: 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # read -ra ver2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@338 -- # local 'op=<' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@340 -- # ver1_l=2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@341 -- # ver2_l=1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@344 -- # case "$op" in 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@345 -- # : 1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # decimal 1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # ver1[v]=1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # decimal 2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # ver2[v]=2 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # return 0 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:07.962 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.962 --rc genhtml_branch_coverage=1 00:24:07.962 --rc genhtml_function_coverage=1 00:24:07.962 --rc genhtml_legend=1 00:24:07.962 --rc geninfo_all_blocks=1 00:24:07.962 --rc geninfo_unexecuted_blocks=1 00:24:07.962 00:24:07.962 ' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:07.962 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.962 --rc genhtml_branch_coverage=1 00:24:07.962 --rc genhtml_function_coverage=1 00:24:07.962 --rc genhtml_legend=1 00:24:07.962 --rc geninfo_all_blocks=1 00:24:07.962 --rc geninfo_unexecuted_blocks=1 00:24:07.962 00:24:07.962 ' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:07.962 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.962 --rc genhtml_branch_coverage=1 00:24:07.962 --rc genhtml_function_coverage=1 00:24:07.962 --rc genhtml_legend=1 00:24:07.962 --rc geninfo_all_blocks=1 00:24:07.962 --rc geninfo_unexecuted_blocks=1 00:24:07.962 00:24:07.962 ' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:07.962 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:07.962 --rc genhtml_branch_coverage=1 00:24:07.962 --rc genhtml_function_coverage=1 00:24:07.962 --rc genhtml_legend=1 00:24:07.962 --rc geninfo_all_blocks=1 00:24:07.962 --rc geninfo_unexecuted_blocks=1 00:24:07.962 00:24:07.962 ' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # uname -s 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@15 -- # shopt -s extglob 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:07.962 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@5 -- # export PATH 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@51 -- # : 0 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:07.963 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@11 -- # MALLOC_BDEV_SIZE=64 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@13 -- # NVMF_HOST_FIRST_PORT=60000 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@14 -- # NVMF_HOST_SECOND_PORT=60001 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@18 -- # '[' tcp == rdma ']' 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@23 -- # nvmftestinit 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@476 -- # prepare_net_devs 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@438 -- # local -g is_hw=no 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # remove_spdk_ns 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@309 -- # xtrace_disable 00:24:07.963 19:20:20 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # pci_devs=() 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # net_devs=() 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # e810=() 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # local -ga e810 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # x722=() 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # local -ga x722 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # mlx=() 00:24:17.963 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # local -ga mlx 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:24:17.964 Found 0000:31:00.0 (0x8086 - 0x159b) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:24:17.964 Found 0000:31:00.1 (0x8086 - 0x159b) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:24:17.964 Found net devices under 0000:31:00.0: cvl_0_0 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:24:17.964 Found net devices under 0000:31:00.1: cvl_0_1 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # is_hw=yes 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:17.964 19:20:28 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:17.964 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:17.964 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.554 ms 00:24:17.964 00:24:17.964 --- 10.0.0.2 ping statistics --- 00:24:17.964 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:17.964 rtt min/avg/max/mdev = 0.554/0.554/0.554/0.000 ms 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:17.964 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:17.964 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.335 ms 00:24:17.964 00:24:17.964 --- 10.0.0.1 ping statistics --- 00:24:17.964 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:17.964 rtt min/avg/max/mdev = 0.335/0.335/0.335/0.000 ms 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@450 -- # return 0 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@25 -- # nvmfappstart -m 0xE 00:24:17.964 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@509 -- # nvmfpid=3945496 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@510 -- # waitforlisten 3945496 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # '[' -z 3945496 ']' 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:17.965 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 [2024-11-26 19:20:29.180120] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:17.965 [2024-11-26 19:20:29.180187] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:17.965 [2024-11-26 19:20:29.287009] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:24:17.965 [2024-11-26 19:20:29.338631] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:17.965 [2024-11-26 19:20:29.338683] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:17.965 [2024-11-26 19:20:29.338696] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:17.965 [2024-11-26 19:20:29.338703] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:17.965 [2024-11-26 19:20:29.338708] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:17.965 [2024-11-26 19:20:29.340755] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:24:17.965 [2024-11-26 19:20:29.340927] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:24:17.965 [2024-11-26 19:20:29.340928] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@868 -- # return 0 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@732 -- # xtrace_disable 00:24:17.965 19:20:29 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@27 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 [2024-11-26 19:20:30.045567] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@29 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 Malloc0 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@30 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 [2024-11-26 19:20:30.113042] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 [2024-11-26 19:20:30.125013] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 Malloc1 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc1 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4421 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@44 -- # bdevperf_pid=3945578 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@46 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; pap "$testdir/try.txt"; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w write -t 1 -f 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@47 -- # waitforlisten 3945578 /var/tmp/bdevperf.sock 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # '[' -z 3945578 ']' 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:24:17.965 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:17.965 19:20:30 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.535 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:18.535 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@868 -- # return 0 00:24:18.535 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@50 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:24:18.535 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.535 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.796 NVMe0n1 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # grep -c NVMe 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:18.796 1 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@60 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.796 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.796 request: 00:24:18.796 { 00:24:18.796 "name": "NVMe0", 00:24:18.796 "trtype": "tcp", 00:24:18.796 "traddr": "10.0.0.2", 00:24:18.796 "adrfam": "ipv4", 00:24:18.796 "trsvcid": "4420", 00:24:18.797 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:24:18.797 "hostnqn": "nqn.2021-09-7.io.spdk:00001", 00:24:18.797 "hostaddr": "10.0.0.1", 00:24:18.797 "prchk_reftag": false, 00:24:18.797 "prchk_guard": false, 00:24:18.797 "hdgst": false, 00:24:18.797 "ddgst": false, 00:24:18.797 "allow_unrecognized_csi": false, 00:24:18.797 "method": "bdev_nvme_attach_controller", 00:24:18.797 "req_id": 1 00:24:18.797 } 00:24:18.797 Got JSON-RPC error response 00:24:18.797 response: 00:24:18.797 { 00:24:18.797 "code": -114, 00:24:18.797 "message": "A controller named NVMe0 already exists with the specified network path" 00:24:18.797 } 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@65 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.797 request: 00:24:18.797 { 00:24:18.797 "name": "NVMe0", 00:24:18.797 "trtype": "tcp", 00:24:18.797 "traddr": "10.0.0.2", 00:24:18.797 "adrfam": "ipv4", 00:24:18.797 "trsvcid": "4420", 00:24:18.797 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:24:18.797 "hostaddr": "10.0.0.1", 00:24:18.797 "prchk_reftag": false, 00:24:18.797 "prchk_guard": false, 00:24:18.797 "hdgst": false, 00:24:18.797 "ddgst": false, 00:24:18.797 "allow_unrecognized_csi": false, 00:24:18.797 "method": "bdev_nvme_attach_controller", 00:24:18.797 "req_id": 1 00:24:18.797 } 00:24:18.797 Got JSON-RPC error response 00:24:18.797 response: 00:24:18.797 { 00:24:18.797 "code": -114, 00:24:18.797 "message": "A controller named NVMe0 already exists with the specified network path" 00:24:18.797 } 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@69 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.797 request: 00:24:18.797 { 00:24:18.797 "name": "NVMe0", 00:24:18.797 "trtype": "tcp", 00:24:18.797 "traddr": "10.0.0.2", 00:24:18.797 "adrfam": "ipv4", 00:24:18.797 "trsvcid": "4420", 00:24:18.797 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:24:18.797 "hostaddr": "10.0.0.1", 00:24:18.797 "prchk_reftag": false, 00:24:18.797 "prchk_guard": false, 00:24:18.797 "hdgst": false, 00:24:18.797 "ddgst": false, 00:24:18.797 "multipath": "disable", 00:24:18.797 "allow_unrecognized_csi": false, 00:24:18.797 "method": "bdev_nvme_attach_controller", 00:24:18.797 "req_id": 1 00:24:18.797 } 00:24:18.797 Got JSON-RPC error response 00:24:18.797 response: 00:24:18.797 { 00:24:18.797 "code": -114, 00:24:18.797 "message": "A controller named NVMe0 already exists and multipath is disabled" 00:24:18.797 } 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@74 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.797 request: 00:24:18.797 { 00:24:18.797 "name": "NVMe0", 00:24:18.797 "trtype": "tcp", 00:24:18.797 "traddr": "10.0.0.2", 00:24:18.797 "adrfam": "ipv4", 00:24:18.797 "trsvcid": "4420", 00:24:18.797 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:24:18.797 "hostaddr": "10.0.0.1", 00:24:18.797 "prchk_reftag": false, 00:24:18.797 "prchk_guard": false, 00:24:18.797 "hdgst": false, 00:24:18.797 "ddgst": false, 00:24:18.797 "multipath": "failover", 00:24:18.797 "allow_unrecognized_csi": false, 00:24:18.797 "method": "bdev_nvme_attach_controller", 00:24:18.797 "req_id": 1 00:24:18.797 } 00:24:18.797 Got JSON-RPC error response 00:24:18.797 response: 00:24:18.797 { 00:24:18.797 "code": -114, 00:24:18.797 "message": "A controller named NVMe0 already exists with the specified network path" 00:24:18.797 } 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@79 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.797 NVMe0n1 00:24:18.797 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:18.798 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@83 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:24:18.798 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.798 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:18.798 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:18.798 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@87 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe1 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:24:18.798 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.798 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:19.057 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # grep -c NVMe 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # '[' 2 '!=' 2 ']' 00:24:19.057 19:20:31 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:24:20.440 { 00:24:20.440 "results": [ 00:24:20.440 { 00:24:20.440 "job": "NVMe0n1", 00:24:20.440 "core_mask": "0x1", 00:24:20.440 "workload": "write", 00:24:20.440 "status": "finished", 00:24:20.440 "queue_depth": 128, 00:24:20.440 "io_size": 4096, 00:24:20.440 "runtime": 1.006045, 00:24:20.440 "iops": 28310.860846184813, 00:24:20.440 "mibps": 110.58930018040942, 00:24:20.440 "io_failed": 0, 00:24:20.440 "io_timeout": 0, 00:24:20.440 "avg_latency_us": 4507.484152330127, 00:24:20.440 "min_latency_us": 2102.6133333333332, 00:24:20.440 "max_latency_us": 9994.24 00:24:20.440 } 00:24:20.440 ], 00:24:20.440 "core_count": 1 00:24:20.440 } 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@98 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe1 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@100 -- # [[ -n '' ]] 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@116 -- # killprocess 3945578 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # '[' -z 3945578 ']' 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@958 -- # kill -0 3945578 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # uname 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3945578 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3945578' 00:24:20.440 killing process with pid 3945578 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@973 -- # kill 3945578 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@978 -- # wait 3945578 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@118 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@119 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@123 -- # pap /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1599 -- # read -r file 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1598 -- # find /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt -type f 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1598 -- # sort -u 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1600 -- # cat 00:24:20.440 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:24:20.440 [2024-11-26 19:20:30.248729] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:20.440 [2024-11-26 19:20:30.248781] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3945578 ] 00:24:20.440 [2024-11-26 19:20:30.325986] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:20.440 [2024-11-26 19:20:30.362443] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:20.440 [2024-11-26 19:20:31.596912] bdev.c:4926:bdev_name_add: *ERROR*: Bdev name 422f0f34-5a23-4c7f-bd72-2a7687b473c8 already exists 00:24:20.440 [2024-11-26 19:20:31.596941] bdev.c:8146:bdev_register: *ERROR*: Unable to add uuid:422f0f34-5a23-4c7f-bd72-2a7687b473c8 alias for bdev NVMe1n1 00:24:20.440 [2024-11-26 19:20:31.596951] bdev_nvme.c:4659:nvme_bdev_create: *ERROR*: spdk_bdev_register() failed 00:24:20.440 Running I/O for 1 seconds... 00:24:20.440 28306.00 IOPS, 110.57 MiB/s 00:24:20.440 Latency(us) 00:24:20.440 [2024-11-26T18:20:33.065Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:20.440 Job: NVMe0n1 (Core Mask 0x1, workload: write, depth: 128, IO size: 4096) 00:24:20.440 NVMe0n1 : 1.01 28310.86 110.59 0.00 0.00 4507.48 2102.61 9994.24 00:24:20.440 [2024-11-26T18:20:33.065Z] =================================================================================================================== 00:24:20.440 [2024-11-26T18:20:33.065Z] Total : 28310.86 110.59 0.00 0.00 4507.48 2102.61 9994.24 00:24:20.440 Received shutdown signal, test time was about 1.000000 seconds 00:24:20.440 00:24:20.440 Latency(us) 00:24:20.440 [2024-11-26T18:20:33.065Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:20.440 [2024-11-26T18:20:33.065Z] =================================================================================================================== 00:24:20.440 [2024-11-26T18:20:33.065Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:20.440 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1605 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1599 -- # read -r file 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@124 -- # nvmftestfini 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@516 -- # nvmfcleanup 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@121 -- # sync 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@124 -- # set +e 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:20.440 19:20:32 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:20.440 rmmod nvme_tcp 00:24:20.440 rmmod nvme_fabrics 00:24:20.440 rmmod nvme_keyring 00:24:20.440 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@128 -- # set -e 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@129 -- # return 0 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@517 -- # '[' -n 3945496 ']' 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@518 -- # killprocess 3945496 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # '[' -z 3945496 ']' 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@958 -- # kill -0 3945496 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # uname 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3945496 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3945496' 00:24:20.701 killing process with pid 3945496 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@973 -- # kill 3945496 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@978 -- # wait 3945496 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@297 -- # iptr 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@791 -- # iptables-save 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@791 -- # iptables-restore 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:20.701 19:20:33 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:23.247 00:24:23.247 real 0m15.020s 00:24:23.247 user 0m17.344s 00:24:23.247 sys 0m7.223s 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:24:23.247 ************************************ 00:24:23.247 END TEST nvmf_multicontroller 00:24:23.247 ************************************ 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@17 -- # run_test nvmf_aer /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:24:23.247 ************************************ 00:24:23.247 START TEST nvmf_aer 00:24:23.247 ************************************ 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:24:23.247 * Looking for test storage... 00:24:23.247 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1693 -- # lcov --version 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # IFS=.-: 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # read -ra ver1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # IFS=.-: 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # read -ra ver2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@338 -- # local 'op=<' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@340 -- # ver1_l=2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@341 -- # ver2_l=1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@344 -- # case "$op" in 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@345 -- # : 1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # decimal 1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # ver1[v]=1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # decimal 2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # ver2[v]=2 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # return 0 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:23.247 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:23.247 --rc genhtml_branch_coverage=1 00:24:23.247 --rc genhtml_function_coverage=1 00:24:23.247 --rc genhtml_legend=1 00:24:23.247 --rc geninfo_all_blocks=1 00:24:23.247 --rc geninfo_unexecuted_blocks=1 00:24:23.247 00:24:23.247 ' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:23.247 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:23.247 --rc genhtml_branch_coverage=1 00:24:23.247 --rc genhtml_function_coverage=1 00:24:23.247 --rc genhtml_legend=1 00:24:23.247 --rc geninfo_all_blocks=1 00:24:23.247 --rc geninfo_unexecuted_blocks=1 00:24:23.247 00:24:23.247 ' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:23.247 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:23.247 --rc genhtml_branch_coverage=1 00:24:23.247 --rc genhtml_function_coverage=1 00:24:23.247 --rc genhtml_legend=1 00:24:23.247 --rc geninfo_all_blocks=1 00:24:23.247 --rc geninfo_unexecuted_blocks=1 00:24:23.247 00:24:23.247 ' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:23.247 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:23.247 --rc genhtml_branch_coverage=1 00:24:23.247 --rc genhtml_function_coverage=1 00:24:23.247 --rc genhtml_legend=1 00:24:23.247 --rc geninfo_all_blocks=1 00:24:23.247 --rc geninfo_unexecuted_blocks=1 00:24:23.247 00:24:23.247 ' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # uname -s 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@15 -- # shopt -s extglob 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:23.247 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@5 -- # export PATH 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@51 -- # : 0 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:23.248 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@11 -- # nvmftestinit 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@476 -- # prepare_net_devs 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@438 -- # local -g is_hw=no 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # remove_spdk_ns 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@309 -- # xtrace_disable 00:24:23.248 19:20:35 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # pci_devs=() 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # net_devs=() 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # e810=() 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # local -ga e810 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # x722=() 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # local -ga x722 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # mlx=() 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # local -ga mlx 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:24:31.384 Found 0000:31:00.0 (0x8086 - 0x159b) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:24:31.384 Found 0000:31:00.1 (0x8086 - 0x159b) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:24:31.384 Found net devices under 0000:31:00.0: cvl_0_0 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:24:31.384 Found net devices under 0000:31:00.1: cvl_0_1 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # is_hw=yes 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:31.384 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:31.384 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.714 ms 00:24:31.384 00:24:31.384 --- 10.0.0.2 ping statistics --- 00:24:31.384 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:31.384 rtt min/avg/max/mdev = 0.714/0.714/0.714/0.000 ms 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:31.384 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:31.384 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.294 ms 00:24:31.384 00:24:31.384 --- 10.0.0.1 ping statistics --- 00:24:31.384 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:31.384 rtt min/avg/max/mdev = 0.294/0.294/0.294/0.000 ms 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@450 -- # return 0 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:24:31.384 19:20:43 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:24:31.643 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@12 -- # nvmfappstart -m 0xF 00:24:31.643 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:24:31.643 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@509 -- # nvmfpid=3950896 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@510 -- # waitforlisten 3950896 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@835 -- # '[' -z 3950896 ']' 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:31.644 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:31.644 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:31.644 [2024-11-26 19:20:44.082730] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:31.644 [2024-11-26 19:20:44.082796] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:31.644 [2024-11-26 19:20:44.172614] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:24:31.644 [2024-11-26 19:20:44.214446] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:31.644 [2024-11-26 19:20:44.214484] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:31.644 [2024-11-26 19:20:44.214492] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:31.644 [2024-11-26 19:20:44.214503] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:31.644 [2024-11-26 19:20:44.214508] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:31.644 [2024-11-26 19:20:44.216380] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:31.644 [2024-11-26 19:20:44.216498] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:24:31.644 [2024-11-26 19:20:44.216653] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:31.644 [2024-11-26 19:20:44.216653] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@868 -- # return 0 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@732 -- # xtrace_disable 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@14 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.581 [2024-11-26 19:20:44.940369] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@16 -- # rpc_cmd bdev_malloc_create 64 512 --name Malloc0 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.581 Malloc0 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@17 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 2 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@18 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.581 19:20:44 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@19 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.581 [2024-11-26 19:20:45.008286] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@21 -- # rpc_cmd nvmf_get_subsystems 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.581 [ 00:24:32.581 { 00:24:32.581 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:24:32.581 "subtype": "Discovery", 00:24:32.581 "listen_addresses": [], 00:24:32.581 "allow_any_host": true, 00:24:32.581 "hosts": [] 00:24:32.581 }, 00:24:32.581 { 00:24:32.581 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:32.581 "subtype": "NVMe", 00:24:32.581 "listen_addresses": [ 00:24:32.581 { 00:24:32.581 "trtype": "TCP", 00:24:32.581 "adrfam": "IPv4", 00:24:32.581 "traddr": "10.0.0.2", 00:24:32.581 "trsvcid": "4420" 00:24:32.581 } 00:24:32.581 ], 00:24:32.581 "allow_any_host": true, 00:24:32.581 "hosts": [], 00:24:32.581 "serial_number": "SPDK00000000000001", 00:24:32.581 "model_number": "SPDK bdev Controller", 00:24:32.581 "max_namespaces": 2, 00:24:32.581 "min_cntlid": 1, 00:24:32.581 "max_cntlid": 65519, 00:24:32.581 "namespaces": [ 00:24:32.581 { 00:24:32.581 "nsid": 1, 00:24:32.581 "bdev_name": "Malloc0", 00:24:32.581 "name": "Malloc0", 00:24:32.581 "nguid": "6F7B4D85BD6E4699915DC2E0A25A0FA9", 00:24:32.581 "uuid": "6f7b4d85-bd6e-4699-915d-c2e0a25a0fa9" 00:24:32.581 } 00:24:32.581 ] 00:24:32.581 } 00:24:32.581 ] 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@23 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@24 -- # rm -f /tmp/aer_touch_file 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@33 -- # aerpid=3951230 00:24:32.581 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@36 -- # waitforfile /tmp/aer_touch_file 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -n 2 -t /tmp/aer_touch_file 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # local i=0 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1271 -- # '[' 0 -lt 200 ']' 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1272 -- # i=1 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1273 -- # sleep 0.1 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1271 -- # '[' 1 -lt 200 ']' 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1272 -- # i=2 00:24:32.582 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1273 -- # sleep 0.1 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1276 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1280 -- # return 0 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@39 -- # rpc_cmd bdev_malloc_create 64 4096 --name Malloc1 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.841 Malloc1 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@40 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 2 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@41 -- # rpc_cmd nvmf_get_subsystems 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.841 Asynchronous Event Request test 00:24:32.841 Attaching to 10.0.0.2 00:24:32.841 Attached to 10.0.0.2 00:24:32.841 Registering asynchronous event callbacks... 00:24:32.841 Starting namespace attribute notice tests for all controllers... 00:24:32.841 10.0.0.2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:24:32.841 aer_cb - Changed Namespace 00:24:32.841 Cleaning up... 00:24:32.841 [ 00:24:32.841 { 00:24:32.841 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:24:32.841 "subtype": "Discovery", 00:24:32.841 "listen_addresses": [], 00:24:32.841 "allow_any_host": true, 00:24:32.841 "hosts": [] 00:24:32.841 }, 00:24:32.841 { 00:24:32.841 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:32.841 "subtype": "NVMe", 00:24:32.841 "listen_addresses": [ 00:24:32.841 { 00:24:32.841 "trtype": "TCP", 00:24:32.841 "adrfam": "IPv4", 00:24:32.841 "traddr": "10.0.0.2", 00:24:32.841 "trsvcid": "4420" 00:24:32.841 } 00:24:32.841 ], 00:24:32.841 "allow_any_host": true, 00:24:32.841 "hosts": [], 00:24:32.841 "serial_number": "SPDK00000000000001", 00:24:32.841 "model_number": "SPDK bdev Controller", 00:24:32.841 "max_namespaces": 2, 00:24:32.841 "min_cntlid": 1, 00:24:32.841 "max_cntlid": 65519, 00:24:32.841 "namespaces": [ 00:24:32.841 { 00:24:32.841 "nsid": 1, 00:24:32.841 "bdev_name": "Malloc0", 00:24:32.841 "name": "Malloc0", 00:24:32.841 "nguid": "6F7B4D85BD6E4699915DC2E0A25A0FA9", 00:24:32.841 "uuid": "6f7b4d85-bd6e-4699-915d-c2e0a25a0fa9" 00:24:32.841 }, 00:24:32.841 { 00:24:32.841 "nsid": 2, 00:24:32.841 "bdev_name": "Malloc1", 00:24:32.841 "name": "Malloc1", 00:24:32.841 "nguid": "FCDFBED9D6C647C7B9D52F9CF0732461", 00:24:32.841 "uuid": "fcdfbed9-d6c6-47c7-b9d5-2f9cf0732461" 00:24:32.841 } 00:24:32.841 ] 00:24:32.841 } 00:24:32.841 ] 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@43 -- # wait 3951230 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@45 -- # rpc_cmd bdev_malloc_delete Malloc0 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@46 -- # rpc_cmd bdev_malloc_delete Malloc1 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@47 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:32.841 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@49 -- # trap - SIGINT SIGTERM EXIT 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@51 -- # nvmftestfini 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@516 -- # nvmfcleanup 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@121 -- # sync 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@124 -- # set +e 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:32.842 rmmod nvme_tcp 00:24:32.842 rmmod nvme_fabrics 00:24:32.842 rmmod nvme_keyring 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@128 -- # set -e 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@129 -- # return 0 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@517 -- # '[' -n 3950896 ']' 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@518 -- # killprocess 3950896 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@954 -- # '[' -z 3950896 ']' 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@958 -- # kill -0 3950896 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@959 -- # uname 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:32.842 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3950896 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3950896' 00:24:33.101 killing process with pid 3950896 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@973 -- # kill 3950896 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@978 -- # wait 3950896 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@297 -- # iptr 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@791 -- # iptables-save 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@791 -- # iptables-restore 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:33.101 19:20:45 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:35.642 00:24:35.642 real 0m12.288s 00:24:35.642 user 0m8.104s 00:24:35.642 sys 0m6.759s 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:24:35.642 ************************************ 00:24:35.642 END TEST nvmf_aer 00:24:35.642 ************************************ 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@18 -- # run_test nvmf_async_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:24:35.642 ************************************ 00:24:35.642 START TEST nvmf_async_init 00:24:35.642 ************************************ 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:24:35.642 * Looking for test storage... 00:24:35.642 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1693 -- # lcov --version 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # IFS=.-: 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # read -ra ver1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # IFS=.-: 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # read -ra ver2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@338 -- # local 'op=<' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@340 -- # ver1_l=2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@341 -- # ver2_l=1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@344 -- # case "$op" in 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@345 -- # : 1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # decimal 1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # ver1[v]=1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # decimal 2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # ver2[v]=2 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # return 0 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:35.642 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:35.642 --rc genhtml_branch_coverage=1 00:24:35.642 --rc genhtml_function_coverage=1 00:24:35.642 --rc genhtml_legend=1 00:24:35.642 --rc geninfo_all_blocks=1 00:24:35.642 --rc geninfo_unexecuted_blocks=1 00:24:35.642 00:24:35.642 ' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:35.642 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:35.642 --rc genhtml_branch_coverage=1 00:24:35.642 --rc genhtml_function_coverage=1 00:24:35.642 --rc genhtml_legend=1 00:24:35.642 --rc geninfo_all_blocks=1 00:24:35.642 --rc geninfo_unexecuted_blocks=1 00:24:35.642 00:24:35.642 ' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:35.642 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:35.642 --rc genhtml_branch_coverage=1 00:24:35.642 --rc genhtml_function_coverage=1 00:24:35.642 --rc genhtml_legend=1 00:24:35.642 --rc geninfo_all_blocks=1 00:24:35.642 --rc geninfo_unexecuted_blocks=1 00:24:35.642 00:24:35.642 ' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:35.642 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:35.642 --rc genhtml_branch_coverage=1 00:24:35.642 --rc genhtml_function_coverage=1 00:24:35.642 --rc genhtml_legend=1 00:24:35.642 --rc geninfo_all_blocks=1 00:24:35.642 --rc geninfo_unexecuted_blocks=1 00:24:35.642 00:24:35.642 ' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # uname -s 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@15 -- # shopt -s extglob 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:35.642 19:20:47 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:35.642 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.642 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.642 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.642 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@5 -- # export PATH 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@51 -- # : 0 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:35.643 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@13 -- # null_bdev_size=1024 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@14 -- # null_block_size=512 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@15 -- # null_bdev=null0 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@16 -- # nvme_bdev=nvme0 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # uuidgen 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # tr -d - 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # nguid=3a63238e607845db8ffc076a549f1b9e 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@22 -- # nvmftestinit 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@476 -- # prepare_net_devs 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@438 -- # local -g is_hw=no 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # remove_spdk_ns 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@309 -- # xtrace_disable 00:24:35.643 19:20:48 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # pci_devs=() 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # net_devs=() 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # e810=() 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # local -ga e810 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # x722=() 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # local -ga x722 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # mlx=() 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # local -ga mlx 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:24:43.794 Found 0000:31:00.0 (0x8086 - 0x159b) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:24:43.794 Found 0000:31:00.1 (0x8086 - 0x159b) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:24:43.794 Found net devices under 0000:31:00.0: cvl_0_0 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:24:43.794 Found net devices under 0000:31:00.1: cvl_0_1 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # is_hw=yes 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:43.794 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:43.795 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:43.795 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.684 ms 00:24:43.795 00:24:43.795 --- 10.0.0.2 ping statistics --- 00:24:43.795 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:43.795 rtt min/avg/max/mdev = 0.684/0.684/0.684/0.000 ms 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:43.795 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:43.795 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.320 ms 00:24:43.795 00:24:43.795 --- 10.0.0.1 ping statistics --- 00:24:43.795 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:43.795 rtt min/avg/max/mdev = 0.320/0.320/0.320/0.000 ms 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@450 -- # return 0 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@23 -- # nvmfappstart -m 0x1 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:43.795 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@509 -- # nvmfpid=3955934 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@510 -- # waitforlisten 3955934 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@835 -- # '[' -z 3955934 ']' 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:44.055 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:44.055 19:20:56 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.055 [2024-11-26 19:20:56.491102] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:44.055 [2024-11-26 19:20:56.491169] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:44.055 [2024-11-26 19:20:56.584868] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:44.055 [2024-11-26 19:20:56.624512] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:44.055 [2024-11-26 19:20:56.624549] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:44.055 [2024-11-26 19:20:56.624557] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:44.055 [2024-11-26 19:20:56.624563] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:44.055 [2024-11-26 19:20:56.624569] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:44.055 [2024-11-26 19:20:56.625176] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@868 -- # return 0 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@732 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@26 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.996 [2024-11-26 19:20:57.323204] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@27 -- # rpc_cmd bdev_null_create null0 1024 512 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.996 null0 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@28 -- # rpc_cmd bdev_wait_for_examine 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@29 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 -g 3a63238e607845db8ffc076a549f1b9e 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@31 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:44.996 [2024-11-26 19:20:57.383499] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@37 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode0 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:44.996 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.257 nvme0n1 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@41 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.257 [ 00:24:45.257 { 00:24:45.257 "name": "nvme0n1", 00:24:45.257 "aliases": [ 00:24:45.257 "3a63238e-6078-45db-8ffc-076a549f1b9e" 00:24:45.257 ], 00:24:45.257 "product_name": "NVMe disk", 00:24:45.257 "block_size": 512, 00:24:45.257 "num_blocks": 2097152, 00:24:45.257 "uuid": "3a63238e-6078-45db-8ffc-076a549f1b9e", 00:24:45.257 "numa_id": 0, 00:24:45.257 "assigned_rate_limits": { 00:24:45.257 "rw_ios_per_sec": 0, 00:24:45.257 "rw_mbytes_per_sec": 0, 00:24:45.257 "r_mbytes_per_sec": 0, 00:24:45.257 "w_mbytes_per_sec": 0 00:24:45.257 }, 00:24:45.257 "claimed": false, 00:24:45.257 "zoned": false, 00:24:45.257 "supported_io_types": { 00:24:45.257 "read": true, 00:24:45.257 "write": true, 00:24:45.257 "unmap": false, 00:24:45.257 "flush": true, 00:24:45.257 "reset": true, 00:24:45.257 "nvme_admin": true, 00:24:45.257 "nvme_io": true, 00:24:45.257 "nvme_io_md": false, 00:24:45.257 "write_zeroes": true, 00:24:45.257 "zcopy": false, 00:24:45.257 "get_zone_info": false, 00:24:45.257 "zone_management": false, 00:24:45.257 "zone_append": false, 00:24:45.257 "compare": true, 00:24:45.257 "compare_and_write": true, 00:24:45.257 "abort": true, 00:24:45.257 "seek_hole": false, 00:24:45.257 "seek_data": false, 00:24:45.257 "copy": true, 00:24:45.257 "nvme_iov_md": false 00:24:45.257 }, 00:24:45.257 "memory_domains": [ 00:24:45.257 { 00:24:45.257 "dma_device_id": "system", 00:24:45.257 "dma_device_type": 1 00:24:45.257 } 00:24:45.257 ], 00:24:45.257 "driver_specific": { 00:24:45.257 "nvme": [ 00:24:45.257 { 00:24:45.257 "trid": { 00:24:45.257 "trtype": "TCP", 00:24:45.257 "adrfam": "IPv4", 00:24:45.257 "traddr": "10.0.0.2", 00:24:45.257 "trsvcid": "4420", 00:24:45.257 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:24:45.257 }, 00:24:45.257 "ctrlr_data": { 00:24:45.257 "cntlid": 1, 00:24:45.257 "vendor_id": "0x8086", 00:24:45.257 "model_number": "SPDK bdev Controller", 00:24:45.257 "serial_number": "00000000000000000000", 00:24:45.257 "firmware_revision": "25.01", 00:24:45.257 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:24:45.257 "oacs": { 00:24:45.257 "security": 0, 00:24:45.257 "format": 0, 00:24:45.257 "firmware": 0, 00:24:45.257 "ns_manage": 0 00:24:45.257 }, 00:24:45.257 "multi_ctrlr": true, 00:24:45.257 "ana_reporting": false 00:24:45.257 }, 00:24:45.257 "vs": { 00:24:45.257 "nvme_version": "1.3" 00:24:45.257 }, 00:24:45.257 "ns_data": { 00:24:45.257 "id": 1, 00:24:45.257 "can_share": true 00:24:45.257 } 00:24:45.257 } 00:24:45.257 ], 00:24:45.257 "mp_policy": "active_passive" 00:24:45.257 } 00:24:45.257 } 00:24:45.257 ] 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@44 -- # rpc_cmd bdev_nvme_reset_controller nvme0 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.257 [2024-11-26 19:20:57.657665] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:24:45.257 [2024-11-26 19:20:57.657727] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2196dd0 (9): Bad file descriptor 00:24:45.257 [2024-11-26 19:20:57.789957] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 2] Resetting controller successful. 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@47 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.257 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.257 [ 00:24:45.257 { 00:24:45.257 "name": "nvme0n1", 00:24:45.257 "aliases": [ 00:24:45.257 "3a63238e-6078-45db-8ffc-076a549f1b9e" 00:24:45.257 ], 00:24:45.257 "product_name": "NVMe disk", 00:24:45.257 "block_size": 512, 00:24:45.257 "num_blocks": 2097152, 00:24:45.257 "uuid": "3a63238e-6078-45db-8ffc-076a549f1b9e", 00:24:45.257 "numa_id": 0, 00:24:45.257 "assigned_rate_limits": { 00:24:45.257 "rw_ios_per_sec": 0, 00:24:45.257 "rw_mbytes_per_sec": 0, 00:24:45.257 "r_mbytes_per_sec": 0, 00:24:45.257 "w_mbytes_per_sec": 0 00:24:45.257 }, 00:24:45.257 "claimed": false, 00:24:45.257 "zoned": false, 00:24:45.257 "supported_io_types": { 00:24:45.257 "read": true, 00:24:45.257 "write": true, 00:24:45.257 "unmap": false, 00:24:45.257 "flush": true, 00:24:45.257 "reset": true, 00:24:45.257 "nvme_admin": true, 00:24:45.257 "nvme_io": true, 00:24:45.257 "nvme_io_md": false, 00:24:45.257 "write_zeroes": true, 00:24:45.257 "zcopy": false, 00:24:45.257 "get_zone_info": false, 00:24:45.257 "zone_management": false, 00:24:45.257 "zone_append": false, 00:24:45.257 "compare": true, 00:24:45.257 "compare_and_write": true, 00:24:45.257 "abort": true, 00:24:45.257 "seek_hole": false, 00:24:45.257 "seek_data": false, 00:24:45.257 "copy": true, 00:24:45.257 "nvme_iov_md": false 00:24:45.257 }, 00:24:45.257 "memory_domains": [ 00:24:45.257 { 00:24:45.257 "dma_device_id": "system", 00:24:45.257 "dma_device_type": 1 00:24:45.257 } 00:24:45.257 ], 00:24:45.257 "driver_specific": { 00:24:45.257 "nvme": [ 00:24:45.257 { 00:24:45.257 "trid": { 00:24:45.257 "trtype": "TCP", 00:24:45.257 "adrfam": "IPv4", 00:24:45.257 "traddr": "10.0.0.2", 00:24:45.257 "trsvcid": "4420", 00:24:45.257 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:24:45.257 }, 00:24:45.257 "ctrlr_data": { 00:24:45.257 "cntlid": 2, 00:24:45.257 "vendor_id": "0x8086", 00:24:45.257 "model_number": "SPDK bdev Controller", 00:24:45.257 "serial_number": "00000000000000000000", 00:24:45.257 "firmware_revision": "25.01", 00:24:45.258 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:24:45.258 "oacs": { 00:24:45.258 "security": 0, 00:24:45.258 "format": 0, 00:24:45.258 "firmware": 0, 00:24:45.258 "ns_manage": 0 00:24:45.258 }, 00:24:45.258 "multi_ctrlr": true, 00:24:45.258 "ana_reporting": false 00:24:45.258 }, 00:24:45.258 "vs": { 00:24:45.258 "nvme_version": "1.3" 00:24:45.258 }, 00:24:45.258 "ns_data": { 00:24:45.258 "id": 1, 00:24:45.258 "can_share": true 00:24:45.258 } 00:24:45.258 } 00:24:45.258 ], 00:24:45.258 "mp_policy": "active_passive" 00:24:45.258 } 00:24:45.258 } 00:24:45.258 ] 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@50 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # mktemp 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # key_path=/tmp/tmp.2nIrkiLoDk 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@54 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@55 -- # chmod 0600 /tmp/tmp.2nIrkiLoDk 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@56 -- # rpc_cmd keyring_file_add_key key0 /tmp/tmp.2nIrkiLoDk 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@57 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode0 --disable 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@58 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 --secure-channel 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.258 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.258 [2024-11-26 19:20:57.878343] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:24:45.258 [2024-11-26 19:20:57.878453] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@60 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host1 --psk key0 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@66 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4421 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host1 --psk key0 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.519 [2024-11-26 19:20:57.902424] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:24:45.519 nvme0n1 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@70 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.519 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.519 [ 00:24:45.519 { 00:24:45.519 "name": "nvme0n1", 00:24:45.519 "aliases": [ 00:24:45.519 "3a63238e-6078-45db-8ffc-076a549f1b9e" 00:24:45.519 ], 00:24:45.519 "product_name": "NVMe disk", 00:24:45.519 "block_size": 512, 00:24:45.519 "num_blocks": 2097152, 00:24:45.519 "uuid": "3a63238e-6078-45db-8ffc-076a549f1b9e", 00:24:45.519 "numa_id": 0, 00:24:45.519 "assigned_rate_limits": { 00:24:45.519 "rw_ios_per_sec": 0, 00:24:45.519 "rw_mbytes_per_sec": 0, 00:24:45.519 "r_mbytes_per_sec": 0, 00:24:45.519 "w_mbytes_per_sec": 0 00:24:45.519 }, 00:24:45.519 "claimed": false, 00:24:45.519 "zoned": false, 00:24:45.519 "supported_io_types": { 00:24:45.519 "read": true, 00:24:45.519 "write": true, 00:24:45.519 "unmap": false, 00:24:45.519 "flush": true, 00:24:45.519 "reset": true, 00:24:45.519 "nvme_admin": true, 00:24:45.519 "nvme_io": true, 00:24:45.519 "nvme_io_md": false, 00:24:45.519 "write_zeroes": true, 00:24:45.519 "zcopy": false, 00:24:45.519 "get_zone_info": false, 00:24:45.519 "zone_management": false, 00:24:45.519 "zone_append": false, 00:24:45.519 "compare": true, 00:24:45.519 "compare_and_write": true, 00:24:45.519 "abort": true, 00:24:45.519 "seek_hole": false, 00:24:45.519 "seek_data": false, 00:24:45.519 "copy": true, 00:24:45.519 "nvme_iov_md": false 00:24:45.519 }, 00:24:45.519 "memory_domains": [ 00:24:45.519 { 00:24:45.519 "dma_device_id": "system", 00:24:45.519 "dma_device_type": 1 00:24:45.519 } 00:24:45.519 ], 00:24:45.519 "driver_specific": { 00:24:45.519 "nvme": [ 00:24:45.519 { 00:24:45.519 "trid": { 00:24:45.519 "trtype": "TCP", 00:24:45.519 "adrfam": "IPv4", 00:24:45.519 "traddr": "10.0.0.2", 00:24:45.519 "trsvcid": "4421", 00:24:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:24:45.519 }, 00:24:45.519 "ctrlr_data": { 00:24:45.519 "cntlid": 3, 00:24:45.519 "vendor_id": "0x8086", 00:24:45.519 "model_number": "SPDK bdev Controller", 00:24:45.519 "serial_number": "00000000000000000000", 00:24:45.519 "firmware_revision": "25.01", 00:24:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:24:45.519 "oacs": { 00:24:45.519 "security": 0, 00:24:45.519 "format": 0, 00:24:45.519 "firmware": 0, 00:24:45.519 "ns_manage": 0 00:24:45.519 }, 00:24:45.519 "multi_ctrlr": true, 00:24:45.519 "ana_reporting": false 00:24:45.519 }, 00:24:45.519 "vs": { 00:24:45.519 "nvme_version": "1.3" 00:24:45.519 }, 00:24:45.519 "ns_data": { 00:24:45.519 "id": 1, 00:24:45.519 "can_share": true 00:24:45.519 } 00:24:45.519 } 00:24:45.519 ], 00:24:45.519 "mp_policy": "active_passive" 00:24:45.519 } 00:24:45.519 } 00:24:45.519 ] 00:24:45.520 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.520 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@73 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:45.520 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.520 19:20:57 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@76 -- # rm -f /tmp/tmp.2nIrkiLoDk 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@78 -- # trap - SIGINT SIGTERM EXIT 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@79 -- # nvmftestfini 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@516 -- # nvmfcleanup 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@121 -- # sync 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@124 -- # set +e 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:45.520 rmmod nvme_tcp 00:24:45.520 rmmod nvme_fabrics 00:24:45.520 rmmod nvme_keyring 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@128 -- # set -e 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@129 -- # return 0 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@517 -- # '[' -n 3955934 ']' 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@518 -- # killprocess 3955934 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@954 -- # '[' -z 3955934 ']' 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@958 -- # kill -0 3955934 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@959 -- # uname 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:45.520 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3955934 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3955934' 00:24:45.780 killing process with pid 3955934 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@973 -- # kill 3955934 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@978 -- # wait 3955934 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@297 -- # iptr 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@791 -- # iptables-save 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:24:45.780 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@791 -- # iptables-restore 00:24:45.781 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:45.781 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:45.781 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:45.781 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:45.781 19:20:58 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:48.326 00:24:48.326 real 0m12.583s 00:24:48.326 user 0m4.455s 00:24:48.326 sys 0m6.659s 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:24:48.326 ************************************ 00:24:48.326 END TEST nvmf_async_init 00:24:48.326 ************************************ 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@19 -- # run_test dma /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:24:48.326 ************************************ 00:24:48.326 START TEST dma 00:24:48.326 ************************************ 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:24:48.326 * Looking for test storage... 00:24:48.326 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1693 -- # lcov --version 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # IFS=.-: 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # read -ra ver1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # IFS=.-: 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # read -ra ver2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@338 -- # local 'op=<' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@340 -- # ver1_l=2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@341 -- # ver2_l=1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@344 -- # case "$op" in 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@345 -- # : 1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # decimal 1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # ver1[v]=1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # decimal 2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # ver2[v]=2 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # return 0 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:48.326 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.326 --rc genhtml_branch_coverage=1 00:24:48.326 --rc genhtml_function_coverage=1 00:24:48.326 --rc genhtml_legend=1 00:24:48.326 --rc geninfo_all_blocks=1 00:24:48.326 --rc geninfo_unexecuted_blocks=1 00:24:48.326 00:24:48.326 ' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:48.326 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.326 --rc genhtml_branch_coverage=1 00:24:48.326 --rc genhtml_function_coverage=1 00:24:48.326 --rc genhtml_legend=1 00:24:48.326 --rc geninfo_all_blocks=1 00:24:48.326 --rc geninfo_unexecuted_blocks=1 00:24:48.326 00:24:48.326 ' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:48.326 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.326 --rc genhtml_branch_coverage=1 00:24:48.326 --rc genhtml_function_coverage=1 00:24:48.326 --rc genhtml_legend=1 00:24:48.326 --rc geninfo_all_blocks=1 00:24:48.326 --rc geninfo_unexecuted_blocks=1 00:24:48.326 00:24:48.326 ' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:48.326 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.326 --rc genhtml_branch_coverage=1 00:24:48.326 --rc genhtml_function_coverage=1 00:24:48.326 --rc genhtml_legend=1 00:24:48.326 --rc geninfo_all_blocks=1 00:24:48.326 --rc geninfo_unexecuted_blocks=1 00:24:48.326 00:24:48.326 ' 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- host/dma.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # uname -s 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:48.326 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@15 -- # shopt -s extglob 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- paths/export.sh@5 -- # export PATH 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@51 -- # : 0 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:48.327 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- host/dma.sh@12 -- # '[' tcp '!=' rdma ']' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- host/dma.sh@13 -- # exit 0 00:24:48.327 00:24:48.327 real 0m0.238s 00:24:48.327 user 0m0.140s 00:24:48.327 sys 0m0.113s 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@10 -- # set +x 00:24:48.327 ************************************ 00:24:48.327 END TEST dma 00:24:48.327 ************************************ 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@22 -- # run_test nvmf_identify /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:24:48.327 ************************************ 00:24:48.327 START TEST nvmf_identify 00:24:48.327 ************************************ 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:24:48.327 * Looking for test storage... 00:24:48.327 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1693 -- # lcov --version 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # IFS=.-: 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # read -ra ver1 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # IFS=.-: 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # read -ra ver2 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@338 -- # local 'op=<' 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@340 -- # ver1_l=2 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@341 -- # ver2_l=1 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@344 -- # case "$op" in 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@345 -- # : 1 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # decimal 1 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=1 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:48.327 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 1 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # ver1[v]=1 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # decimal 2 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=2 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 2 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # ver2[v]=2 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # return 0 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:48.589 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.589 --rc genhtml_branch_coverage=1 00:24:48.589 --rc genhtml_function_coverage=1 00:24:48.589 --rc genhtml_legend=1 00:24:48.589 --rc geninfo_all_blocks=1 00:24:48.589 --rc geninfo_unexecuted_blocks=1 00:24:48.589 00:24:48.589 ' 00:24:48.589 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:48.589 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.589 --rc genhtml_branch_coverage=1 00:24:48.589 --rc genhtml_function_coverage=1 00:24:48.589 --rc genhtml_legend=1 00:24:48.589 --rc geninfo_all_blocks=1 00:24:48.589 --rc geninfo_unexecuted_blocks=1 00:24:48.589 00:24:48.589 ' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:48.590 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.590 --rc genhtml_branch_coverage=1 00:24:48.590 --rc genhtml_function_coverage=1 00:24:48.590 --rc genhtml_legend=1 00:24:48.590 --rc geninfo_all_blocks=1 00:24:48.590 --rc geninfo_unexecuted_blocks=1 00:24:48.590 00:24:48.590 ' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:48.590 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:48.590 --rc genhtml_branch_coverage=1 00:24:48.590 --rc genhtml_function_coverage=1 00:24:48.590 --rc genhtml_legend=1 00:24:48.590 --rc geninfo_all_blocks=1 00:24:48.590 --rc geninfo_unexecuted_blocks=1 00:24:48.590 00:24:48.590 ' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # uname -s 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@15 -- # shopt -s extglob 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@5 -- # export PATH 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@51 -- # : 0 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:48.590 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@11 -- # MALLOC_BDEV_SIZE=64 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@14 -- # nvmftestinit 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@476 -- # prepare_net_devs 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@438 -- # local -g is_hw=no 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # remove_spdk_ns 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:48.590 19:21:00 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:48.590 19:21:01 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:24:48.590 19:21:01 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:24:48.590 19:21:01 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@309 -- # xtrace_disable 00:24:48.590 19:21:01 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # pci_devs=() 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # net_devs=() 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # e810=() 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # local -ga e810 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # x722=() 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # local -ga x722 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # mlx=() 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # local -ga mlx 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:56.732 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:24:56.733 Found 0000:31:00.0 (0x8086 - 0x159b) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:24:56.733 Found 0000:31:00.1 (0x8086 - 0x159b) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:24:56.733 Found net devices under 0000:31:00.0: cvl_0_0 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@418 -- # [[ up == up ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:24:56.733 Found net devices under 0000:31:00.1: cvl_0_1 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # is_hw=yes 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:56.733 19:21:08 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:56.733 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:56.733 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.634 ms 00:24:56.733 00:24:56.733 --- 10.0.0.2 ping statistics --- 00:24:56.733 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:56.733 rtt min/avg/max/mdev = 0.634/0.634/0.634/0.000 ms 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:56.733 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:56.733 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.311 ms 00:24:56.733 00:24:56.733 --- 10.0.0.1 ping statistics --- 00:24:56.733 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:56.733 rtt min/avg/max/mdev = 0.311/0.311/0.311/0.000 ms 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@450 -- # return 0 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@16 -- # timing_enter start_nvmf_tgt 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@19 -- # nvmfpid=3961354 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@21 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@18 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@23 -- # waitforlisten 3961354 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@835 -- # '[' -z 3961354 ']' 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:56.733 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:56.733 19:21:09 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:56.733 [2024-11-26 19:21:09.303735] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:56.733 [2024-11-26 19:21:09.303819] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:56.994 [2024-11-26 19:21:09.395102] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:24:56.994 [2024-11-26 19:21:09.438133] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:56.994 [2024-11-26 19:21:09.438170] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:56.994 [2024-11-26 19:21:09.438178] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:56.994 [2024-11-26 19:21:09.438185] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:56.994 [2024-11-26 19:21:09.438191] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:56.994 [2024-11-26 19:21:09.439780] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:56.994 [2024-11-26 19:21:09.439904] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:24:56.994 [2024-11-26 19:21:09.440054] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:56.994 [2024-11-26 19:21:09.440054] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@868 -- # return 0 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@24 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.565 [2024-11-26 19:21:10.103851] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@25 -- # timing_exit start_nvmf_tgt 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@732 -- # xtrace_disable 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@27 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.565 Malloc0 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.565 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 --nguid ABCDEF0123456789ABCDEF0123456789 --eui64 ABCDEF0123456789 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.827 [2024-11-26 19:21:10.215036] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@35 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@37 -- # rpc_cmd nvmf_get_subsystems 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.827 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:57.827 [ 00:24:57.827 { 00:24:57.827 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:24:57.827 "subtype": "Discovery", 00:24:57.827 "listen_addresses": [ 00:24:57.827 { 00:24:57.827 "trtype": "TCP", 00:24:57.827 "adrfam": "IPv4", 00:24:57.827 "traddr": "10.0.0.2", 00:24:57.827 "trsvcid": "4420" 00:24:57.827 } 00:24:57.827 ], 00:24:57.827 "allow_any_host": true, 00:24:57.827 "hosts": [] 00:24:57.827 }, 00:24:57.827 { 00:24:57.827 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:57.827 "subtype": "NVMe", 00:24:57.827 "listen_addresses": [ 00:24:57.827 { 00:24:57.827 "trtype": "TCP", 00:24:57.827 "adrfam": "IPv4", 00:24:57.828 "traddr": "10.0.0.2", 00:24:57.828 "trsvcid": "4420" 00:24:57.828 } 00:24:57.828 ], 00:24:57.828 "allow_any_host": true, 00:24:57.828 "hosts": [], 00:24:57.828 "serial_number": "SPDK00000000000001", 00:24:57.828 "model_number": "SPDK bdev Controller", 00:24:57.828 "max_namespaces": 32, 00:24:57.828 "min_cntlid": 1, 00:24:57.828 "max_cntlid": 65519, 00:24:57.828 "namespaces": [ 00:24:57.828 { 00:24:57.828 "nsid": 1, 00:24:57.828 "bdev_name": "Malloc0", 00:24:57.828 "name": "Malloc0", 00:24:57.828 "nguid": "ABCDEF0123456789ABCDEF0123456789", 00:24:57.828 "eui64": "ABCDEF0123456789", 00:24:57.828 "uuid": "94e86b4e-3725-47e6-ad26-4d43833a09c9" 00:24:57.828 } 00:24:57.828 ] 00:24:57.828 } 00:24:57.828 ] 00:24:57.828 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.828 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' -L all 00:24:57.828 [2024-11-26 19:21:10.277574] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:57.828 [2024-11-26 19:21:10.277617] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3961831 ] 00:24:57.828 [2024-11-26 19:21:10.330058] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 0] setting state to connect adminq (no timeout) 00:24:57.828 [2024-11-26 19:21:10.330113] nvme_tcp.c:2238:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:24:57.828 [2024-11-26 19:21:10.330119] nvme_tcp.c:2242:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:24:57.828 [2024-11-26 19:21:10.330135] nvme_tcp.c:2263:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:24:57.828 [2024-11-26 19:21:10.330145] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:24:57.828 [2024-11-26 19:21:10.334172] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 0] setting state to wait for connect adminq (no timeout) 00:24:57.828 [2024-11-26 19:21:10.334221] nvme_tcp.c:1455:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x1af2550 0 00:24:57.828 [2024-11-26 19:21:10.341874] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:24:57.828 [2024-11-26 19:21:10.341886] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:24:57.828 [2024-11-26 19:21:10.341891] nvme_tcp.c:1501:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:24:57.828 [2024-11-26 19:21:10.341896] nvme_tcp.c:1502:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:24:57.828 [2024-11-26 19:21:10.341930] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.341937] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.341942] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.828 [2024-11-26 19:21:10.341956] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:24:57.828 [2024-11-26 19:21:10.341974] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.828 [2024-11-26 19:21:10.349874] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.828 [2024-11-26 19:21:10.349884] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.828 [2024-11-26 19:21:10.349887] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.349893] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.828 [2024-11-26 19:21:10.349905] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:24:57.828 [2024-11-26 19:21:10.349913] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read vs (no timeout) 00:24:57.828 [2024-11-26 19:21:10.349918] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read vs wait for vs (no timeout) 00:24:57.828 [2024-11-26 19:21:10.349934] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.349938] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.349942] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.828 [2024-11-26 19:21:10.349950] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.828 [2024-11-26 19:21:10.349964] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.828 [2024-11-26 19:21:10.350148] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.828 [2024-11-26 19:21:10.350155] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.828 [2024-11-26 19:21:10.350158] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350162] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.828 [2024-11-26 19:21:10.350170] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read cap (no timeout) 00:24:57.828 [2024-11-26 19:21:10.350177] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read cap wait for cap (no timeout) 00:24:57.828 [2024-11-26 19:21:10.350184] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350188] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350192] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.828 [2024-11-26 19:21:10.350199] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.828 [2024-11-26 19:21:10.350209] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.828 [2024-11-26 19:21:10.350369] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.828 [2024-11-26 19:21:10.350376] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.828 [2024-11-26 19:21:10.350382] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350386] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.828 [2024-11-26 19:21:10.350392] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to check en (no timeout) 00:24:57.828 [2024-11-26 19:21:10.350400] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to check en wait for cc (timeout 15000 ms) 00:24:57.828 [2024-11-26 19:21:10.350407] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350410] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350414] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.828 [2024-11-26 19:21:10.350421] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.828 [2024-11-26 19:21:10.350431] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.828 [2024-11-26 19:21:10.350625] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.828 [2024-11-26 19:21:10.350631] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.828 [2024-11-26 19:21:10.350635] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350638] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.828 [2024-11-26 19:21:10.350643] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:24:57.828 [2024-11-26 19:21:10.350653] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350657] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350660] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.828 [2024-11-26 19:21:10.350667] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.828 [2024-11-26 19:21:10.350677] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.828 [2024-11-26 19:21:10.350834] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.828 [2024-11-26 19:21:10.350840] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.828 [2024-11-26 19:21:10.350844] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.350848] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.828 [2024-11-26 19:21:10.350852] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CC.EN = 0 && CSTS.RDY = 0 00:24:57.828 [2024-11-26 19:21:10.350857] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to controller is disabled (timeout 15000 ms) 00:24:57.828 [2024-11-26 19:21:10.350872] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:24:57.828 [2024-11-26 19:21:10.350984] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Setting CC.EN = 1 00:24:57.828 [2024-11-26 19:21:10.350988] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:24:57.828 [2024-11-26 19:21:10.350997] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.351001] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.351005] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.828 [2024-11-26 19:21:10.351011] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.828 [2024-11-26 19:21:10.351024] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.828 [2024-11-26 19:21:10.351188] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.828 [2024-11-26 19:21:10.351195] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.828 [2024-11-26 19:21:10.351198] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.351202] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.828 [2024-11-26 19:21:10.351207] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:24:57.828 [2024-11-26 19:21:10.351216] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.351220] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.828 [2024-11-26 19:21:10.351223] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.828 [2024-11-26 19:21:10.351230] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.828 [2024-11-26 19:21:10.351240] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.828 [2024-11-26 19:21:10.351400] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.828 [2024-11-26 19:21:10.351406] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.829 [2024-11-26 19:21:10.351410] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.351414] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.829 [2024-11-26 19:21:10.351418] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:24:57.829 [2024-11-26 19:21:10.351424] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to reset admin queue (timeout 30000 ms) 00:24:57.829 [2024-11-26 19:21:10.351431] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to identify controller (no timeout) 00:24:57.829 [2024-11-26 19:21:10.351445] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for identify controller (timeout 30000 ms) 00:24:57.829 [2024-11-26 19:21:10.351454] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.351458] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.351465] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.829 [2024-11-26 19:21:10.351475] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.829 [2024-11-26 19:21:10.351665] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:57.829 [2024-11-26 19:21:10.351671] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:57.829 [2024-11-26 19:21:10.351675] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.351679] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1af2550): datao=0, datal=4096, cccid=0 00:24:57.829 [2024-11-26 19:21:10.351684] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1b54100) on tqpair(0x1af2550): expected_datao=0, payload_size=4096 00:24:57.829 [2024-11-26 19:21:10.351688] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.351715] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.351719] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392000] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.829 [2024-11-26 19:21:10.392010] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.829 [2024-11-26 19:21:10.392014] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392021] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.829 [2024-11-26 19:21:10.392030] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] transport max_xfer_size 4294967295 00:24:57.829 [2024-11-26 19:21:10.392035] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] MDTS max_xfer_size 131072 00:24:57.829 [2024-11-26 19:21:10.392040] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CNTLID 0x0001 00:24:57.829 [2024-11-26 19:21:10.392045] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] transport max_sges 16 00:24:57.829 [2024-11-26 19:21:10.392050] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] fuses compare and write: 1 00:24:57.829 [2024-11-26 19:21:10.392055] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to configure AER (timeout 30000 ms) 00:24:57.829 [2024-11-26 19:21:10.392064] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for configure aer (timeout 30000 ms) 00:24:57.829 [2024-11-26 19:21:10.392071] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392075] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392079] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.392087] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:24:57.829 [2024-11-26 19:21:10.392099] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.829 [2024-11-26 19:21:10.392312] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.829 [2024-11-26 19:21:10.392319] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.829 [2024-11-26 19:21:10.392322] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392326] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:57.829 [2024-11-26 19:21:10.392334] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392338] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392341] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.392348] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:57.829 [2024-11-26 19:21:10.392354] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392358] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392362] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.392368] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:57.829 [2024-11-26 19:21:10.392374] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392378] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392382] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.392387] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:57.829 [2024-11-26 19:21:10.392394] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392398] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392401] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.392407] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:57.829 [2024-11-26 19:21:10.392414] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to set keep alive timeout (timeout 30000 ms) 00:24:57.829 [2024-11-26 19:21:10.392425] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:24:57.829 [2024-11-26 19:21:10.392432] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392436] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.392443] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.829 [2024-11-26 19:21:10.392455] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54100, cid 0, qid 0 00:24:57.829 [2024-11-26 19:21:10.392460] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54280, cid 1, qid 0 00:24:57.829 [2024-11-26 19:21:10.392465] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54400, cid 2, qid 0 00:24:57.829 [2024-11-26 19:21:10.392470] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:57.829 [2024-11-26 19:21:10.392475] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54700, cid 4, qid 0 00:24:57.829 [2024-11-26 19:21:10.392706] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.829 [2024-11-26 19:21:10.392713] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.829 [2024-11-26 19:21:10.392716] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392720] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54700) on tqpair=0x1af2550 00:24:57.829 [2024-11-26 19:21:10.392725] nvme_ctrlr.c:3059:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Sending keep alive every 5000000 us 00:24:57.829 [2024-11-26 19:21:10.392730] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to ready (no timeout) 00:24:57.829 [2024-11-26 19:21:10.392741] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392745] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.392752] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.829 [2024-11-26 19:21:10.392762] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54700, cid 4, qid 0 00:24:57.829 [2024-11-26 19:21:10.392954] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:57.829 [2024-11-26 19:21:10.392962] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:57.829 [2024-11-26 19:21:10.392965] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392969] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1af2550): datao=0, datal=4096, cccid=4 00:24:57.829 [2024-11-26 19:21:10.392974] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1b54700) on tqpair(0x1af2550): expected_datao=0, payload_size=4096 00:24:57.829 [2024-11-26 19:21:10.392978] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392985] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.392989] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.393159] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.829 [2024-11-26 19:21:10.393166] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.829 [2024-11-26 19:21:10.393169] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.393173] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54700) on tqpair=0x1af2550 00:24:57.829 [2024-11-26 19:21:10.393185] nvme_ctrlr.c:4202:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Ctrlr already in ready state 00:24:57.829 [2024-11-26 19:21:10.393211] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.393215] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.393222] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.829 [2024-11-26 19:21:10.393230] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.393233] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.829 [2024-11-26 19:21:10.393237] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1af2550) 00:24:57.829 [2024-11-26 19:21:10.393243] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:24:57.829 [2024-11-26 19:21:10.393257] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54700, cid 4, qid 0 00:24:57.829 [2024-11-26 19:21:10.393262] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54880, cid 5, qid 0 00:24:57.829 [2024-11-26 19:21:10.393494] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:57.829 [2024-11-26 19:21:10.393500] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:57.830 [2024-11-26 19:21:10.393504] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.393507] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1af2550): datao=0, datal=1024, cccid=4 00:24:57.830 [2024-11-26 19:21:10.393512] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1b54700) on tqpair(0x1af2550): expected_datao=0, payload_size=1024 00:24:57.830 [2024-11-26 19:21:10.393516] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.393523] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.393526] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.393532] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.830 [2024-11-26 19:21:10.393538] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.830 [2024-11-26 19:21:10.393542] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.393546] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54880) on tqpair=0x1af2550 00:24:57.830 [2024-11-26 19:21:10.437872] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.830 [2024-11-26 19:21:10.437884] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.830 [2024-11-26 19:21:10.437888] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.437892] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54700) on tqpair=0x1af2550 00:24:57.830 [2024-11-26 19:21:10.437904] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.437908] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1af2550) 00:24:57.830 [2024-11-26 19:21:10.437916] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:02ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.830 [2024-11-26 19:21:10.437932] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54700, cid 4, qid 0 00:24:57.830 [2024-11-26 19:21:10.438130] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:57.830 [2024-11-26 19:21:10.438137] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:57.830 [2024-11-26 19:21:10.438140] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438144] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1af2550): datao=0, datal=3072, cccid=4 00:24:57.830 [2024-11-26 19:21:10.438148] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1b54700) on tqpair(0x1af2550): expected_datao=0, payload_size=3072 00:24:57.830 [2024-11-26 19:21:10.438153] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438159] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438166] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438329] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:57.830 [2024-11-26 19:21:10.438335] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:57.830 [2024-11-26 19:21:10.438338] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438342] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54700) on tqpair=0x1af2550 00:24:57.830 [2024-11-26 19:21:10.438351] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438354] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1af2550) 00:24:57.830 [2024-11-26 19:21:10.438361] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00010070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:57.830 [2024-11-26 19:21:10.438374] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54700, cid 4, qid 0 00:24:57.830 [2024-11-26 19:21:10.438584] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:57.830 [2024-11-26 19:21:10.438591] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:57.830 [2024-11-26 19:21:10.438594] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438598] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1af2550): datao=0, datal=8, cccid=4 00:24:57.830 [2024-11-26 19:21:10.438602] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1b54700) on tqpair(0x1af2550): expected_datao=0, payload_size=8 00:24:57.830 [2024-11-26 19:21:10.438606] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438613] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:57.830 [2024-11-26 19:21:10.438616] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.096 [2024-11-26 19:21:10.479023] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.096 [2024-11-26 19:21:10.479034] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.096 [2024-11-26 19:21:10.479038] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.096 [2024-11-26 19:21:10.479042] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54700) on tqpair=0x1af2550 00:24:58.096 ===================================================== 00:24:58.096 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2014-08.org.nvmexpress.discovery 00:24:58.096 ===================================================== 00:24:58.096 Controller Capabilities/Features 00:24:58.096 ================================ 00:24:58.096 Vendor ID: 0000 00:24:58.096 Subsystem Vendor ID: 0000 00:24:58.096 Serial Number: .................... 00:24:58.096 Model Number: ........................................ 00:24:58.096 Firmware Version: 25.01 00:24:58.096 Recommended Arb Burst: 0 00:24:58.096 IEEE OUI Identifier: 00 00 00 00:24:58.096 Multi-path I/O 00:24:58.096 May have multiple subsystem ports: No 00:24:58.096 May have multiple controllers: No 00:24:58.096 Associated with SR-IOV VF: No 00:24:58.096 Max Data Transfer Size: 131072 00:24:58.096 Max Number of Namespaces: 0 00:24:58.096 Max Number of I/O Queues: 1024 00:24:58.096 NVMe Specification Version (VS): 1.3 00:24:58.096 NVMe Specification Version (Identify): 1.3 00:24:58.096 Maximum Queue Entries: 128 00:24:58.096 Contiguous Queues Required: Yes 00:24:58.096 Arbitration Mechanisms Supported 00:24:58.096 Weighted Round Robin: Not Supported 00:24:58.096 Vendor Specific: Not Supported 00:24:58.096 Reset Timeout: 15000 ms 00:24:58.096 Doorbell Stride: 4 bytes 00:24:58.096 NVM Subsystem Reset: Not Supported 00:24:58.096 Command Sets Supported 00:24:58.096 NVM Command Set: Supported 00:24:58.096 Boot Partition: Not Supported 00:24:58.096 Memory Page Size Minimum: 4096 bytes 00:24:58.096 Memory Page Size Maximum: 4096 bytes 00:24:58.096 Persistent Memory Region: Not Supported 00:24:58.096 Optional Asynchronous Events Supported 00:24:58.096 Namespace Attribute Notices: Not Supported 00:24:58.096 Firmware Activation Notices: Not Supported 00:24:58.096 ANA Change Notices: Not Supported 00:24:58.096 PLE Aggregate Log Change Notices: Not Supported 00:24:58.096 LBA Status Info Alert Notices: Not Supported 00:24:58.096 EGE Aggregate Log Change Notices: Not Supported 00:24:58.096 Normal NVM Subsystem Shutdown event: Not Supported 00:24:58.096 Zone Descriptor Change Notices: Not Supported 00:24:58.096 Discovery Log Change Notices: Supported 00:24:58.096 Controller Attributes 00:24:58.096 128-bit Host Identifier: Not Supported 00:24:58.096 Non-Operational Permissive Mode: Not Supported 00:24:58.096 NVM Sets: Not Supported 00:24:58.096 Read Recovery Levels: Not Supported 00:24:58.096 Endurance Groups: Not Supported 00:24:58.096 Predictable Latency Mode: Not Supported 00:24:58.096 Traffic Based Keep ALive: Not Supported 00:24:58.096 Namespace Granularity: Not Supported 00:24:58.096 SQ Associations: Not Supported 00:24:58.096 UUID List: Not Supported 00:24:58.096 Multi-Domain Subsystem: Not Supported 00:24:58.096 Fixed Capacity Management: Not Supported 00:24:58.096 Variable Capacity Management: Not Supported 00:24:58.096 Delete Endurance Group: Not Supported 00:24:58.096 Delete NVM Set: Not Supported 00:24:58.096 Extended LBA Formats Supported: Not Supported 00:24:58.096 Flexible Data Placement Supported: Not Supported 00:24:58.096 00:24:58.096 Controller Memory Buffer Support 00:24:58.096 ================================ 00:24:58.096 Supported: No 00:24:58.096 00:24:58.096 Persistent Memory Region Support 00:24:58.096 ================================ 00:24:58.096 Supported: No 00:24:58.096 00:24:58.096 Admin Command Set Attributes 00:24:58.096 ============================ 00:24:58.097 Security Send/Receive: Not Supported 00:24:58.097 Format NVM: Not Supported 00:24:58.097 Firmware Activate/Download: Not Supported 00:24:58.097 Namespace Management: Not Supported 00:24:58.097 Device Self-Test: Not Supported 00:24:58.097 Directives: Not Supported 00:24:58.097 NVMe-MI: Not Supported 00:24:58.097 Virtualization Management: Not Supported 00:24:58.097 Doorbell Buffer Config: Not Supported 00:24:58.097 Get LBA Status Capability: Not Supported 00:24:58.097 Command & Feature Lockdown Capability: Not Supported 00:24:58.097 Abort Command Limit: 1 00:24:58.097 Async Event Request Limit: 4 00:24:58.097 Number of Firmware Slots: N/A 00:24:58.097 Firmware Slot 1 Read-Only: N/A 00:24:58.097 Firmware Activation Without Reset: N/A 00:24:58.097 Multiple Update Detection Support: N/A 00:24:58.097 Firmware Update Granularity: No Information Provided 00:24:58.097 Per-Namespace SMART Log: No 00:24:58.097 Asymmetric Namespace Access Log Page: Not Supported 00:24:58.097 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:24:58.097 Command Effects Log Page: Not Supported 00:24:58.097 Get Log Page Extended Data: Supported 00:24:58.097 Telemetry Log Pages: Not Supported 00:24:58.097 Persistent Event Log Pages: Not Supported 00:24:58.097 Supported Log Pages Log Page: May Support 00:24:58.097 Commands Supported & Effects Log Page: Not Supported 00:24:58.097 Feature Identifiers & Effects Log Page:May Support 00:24:58.097 NVMe-MI Commands & Effects Log Page: May Support 00:24:58.097 Data Area 4 for Telemetry Log: Not Supported 00:24:58.097 Error Log Page Entries Supported: 128 00:24:58.097 Keep Alive: Not Supported 00:24:58.097 00:24:58.097 NVM Command Set Attributes 00:24:58.097 ========================== 00:24:58.097 Submission Queue Entry Size 00:24:58.097 Max: 1 00:24:58.097 Min: 1 00:24:58.097 Completion Queue Entry Size 00:24:58.097 Max: 1 00:24:58.097 Min: 1 00:24:58.097 Number of Namespaces: 0 00:24:58.097 Compare Command: Not Supported 00:24:58.097 Write Uncorrectable Command: Not Supported 00:24:58.097 Dataset Management Command: Not Supported 00:24:58.097 Write Zeroes Command: Not Supported 00:24:58.097 Set Features Save Field: Not Supported 00:24:58.097 Reservations: Not Supported 00:24:58.097 Timestamp: Not Supported 00:24:58.097 Copy: Not Supported 00:24:58.097 Volatile Write Cache: Not Present 00:24:58.097 Atomic Write Unit (Normal): 1 00:24:58.097 Atomic Write Unit (PFail): 1 00:24:58.097 Atomic Compare & Write Unit: 1 00:24:58.097 Fused Compare & Write: Supported 00:24:58.097 Scatter-Gather List 00:24:58.097 SGL Command Set: Supported 00:24:58.097 SGL Keyed: Supported 00:24:58.097 SGL Bit Bucket Descriptor: Not Supported 00:24:58.097 SGL Metadata Pointer: Not Supported 00:24:58.097 Oversized SGL: Not Supported 00:24:58.097 SGL Metadata Address: Not Supported 00:24:58.097 SGL Offset: Supported 00:24:58.097 Transport SGL Data Block: Not Supported 00:24:58.097 Replay Protected Memory Block: Not Supported 00:24:58.097 00:24:58.097 Firmware Slot Information 00:24:58.097 ========================= 00:24:58.097 Active slot: 0 00:24:58.097 00:24:58.097 00:24:58.097 Error Log 00:24:58.097 ========= 00:24:58.097 00:24:58.097 Active Namespaces 00:24:58.097 ================= 00:24:58.097 Discovery Log Page 00:24:58.097 ================== 00:24:58.097 Generation Counter: 2 00:24:58.097 Number of Records: 2 00:24:58.097 Record Format: 0 00:24:58.097 00:24:58.097 Discovery Log Entry 0 00:24:58.097 ---------------------- 00:24:58.097 Transport Type: 3 (TCP) 00:24:58.097 Address Family: 1 (IPv4) 00:24:58.097 Subsystem Type: 3 (Current Discovery Subsystem) 00:24:58.097 Entry Flags: 00:24:58.097 Duplicate Returned Information: 1 00:24:58.097 Explicit Persistent Connection Support for Discovery: 1 00:24:58.097 Transport Requirements: 00:24:58.097 Secure Channel: Not Required 00:24:58.097 Port ID: 0 (0x0000) 00:24:58.097 Controller ID: 65535 (0xffff) 00:24:58.097 Admin Max SQ Size: 128 00:24:58.097 Transport Service Identifier: 4420 00:24:58.097 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:24:58.097 Transport Address: 10.0.0.2 00:24:58.097 Discovery Log Entry 1 00:24:58.097 ---------------------- 00:24:58.097 Transport Type: 3 (TCP) 00:24:58.097 Address Family: 1 (IPv4) 00:24:58.097 Subsystem Type: 2 (NVM Subsystem) 00:24:58.097 Entry Flags: 00:24:58.097 Duplicate Returned Information: 0 00:24:58.097 Explicit Persistent Connection Support for Discovery: 0 00:24:58.097 Transport Requirements: 00:24:58.097 Secure Channel: Not Required 00:24:58.097 Port ID: 0 (0x0000) 00:24:58.097 Controller ID: 65535 (0xffff) 00:24:58.097 Admin Max SQ Size: 128 00:24:58.097 Transport Service Identifier: 4420 00:24:58.097 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:cnode1 00:24:58.097 Transport Address: 10.0.0.2 [2024-11-26 19:21:10.479131] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Prepare to destruct SSD 00:24:58.097 [2024-11-26 19:21:10.479142] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54100) on tqpair=0x1af2550 00:24:58.097 [2024-11-26 19:21:10.479148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.097 [2024-11-26 19:21:10.479154] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54280) on tqpair=0x1af2550 00:24:58.097 [2024-11-26 19:21:10.479158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.097 [2024-11-26 19:21:10.479163] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54400) on tqpair=0x1af2550 00:24:58.097 [2024-11-26 19:21:10.479168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.097 [2024-11-26 19:21:10.479173] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.097 [2024-11-26 19:21:10.479177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.097 [2024-11-26 19:21:10.479186] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479190] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479194] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.479201] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.479217] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.479300] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.479307] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.479311] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479314] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.479322] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479325] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479329] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.479336] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.479349] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.479526] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.479533] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.479536] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479540] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.479545] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] RTD3E = 0 us 00:24:58.098 [2024-11-26 19:21:10.479550] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] shutdown timeout = 10000 ms 00:24:58.098 [2024-11-26 19:21:10.479559] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479563] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479567] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.479574] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.479584] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.479739] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.479746] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.479750] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479754] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.479764] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479768] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479771] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.479778] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.479788] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.479970] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.479977] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.479981] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479985] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.479994] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.479998] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480002] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.480011] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.480021] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.480196] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.480202] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.480206] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480210] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.480219] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480223] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480227] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.480234] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.480243] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.480426] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.480433] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.480436] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480440] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.480450] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480454] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480457] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.480464] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.480474] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.480652] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.480658] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.480661] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480665] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.480674] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480678] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.480682] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.480689] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.480699] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.484873] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.484882] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.098 [2024-11-26 19:21:10.484886] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.484889] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.098 [2024-11-26 19:21:10.484899] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.484903] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.098 [2024-11-26 19:21:10.484907] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1af2550) 00:24:58.098 [2024-11-26 19:21:10.484914] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.098 [2024-11-26 19:21:10.484928] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1b54580, cid 3, qid 0 00:24:58.098 [2024-11-26 19:21:10.485106] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.098 [2024-11-26 19:21:10.485113] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.099 [2024-11-26 19:21:10.485117] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.485121] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1b54580) on tqpair=0x1af2550 00:24:58.099 [2024-11-26 19:21:10.485128] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] shutdown complete in 5 milliseconds 00:24:58.099 00:24:58.099 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -L all 00:24:58.099 [2024-11-26 19:21:10.528648] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:24:58.099 [2024-11-26 19:21:10.528697] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3961891 ] 00:24:58.099 [2024-11-26 19:21:10.592025] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 0] setting state to connect adminq (no timeout) 00:24:58.099 [2024-11-26 19:21:10.592073] nvme_tcp.c:2238:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:24:58.099 [2024-11-26 19:21:10.592078] nvme_tcp.c:2242:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:24:58.099 [2024-11-26 19:21:10.592097] nvme_tcp.c:2263:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:24:58.099 [2024-11-26 19:21:10.592106] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:24:58.099 [2024-11-26 19:21:10.596065] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 0] setting state to wait for connect adminq (no timeout) 00:24:58.099 [2024-11-26 19:21:10.596095] nvme_tcp.c:1455:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x162f550 0 00:24:58.099 [2024-11-26 19:21:10.596297] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:24:58.099 [2024-11-26 19:21:10.596305] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:24:58.099 [2024-11-26 19:21:10.596309] nvme_tcp.c:1501:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:24:58.099 [2024-11-26 19:21:10.596312] nvme_tcp.c:1502:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:24:58.099 [2024-11-26 19:21:10.596338] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.596344] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.596348] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.099 [2024-11-26 19:21:10.596359] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:24:58.099 [2024-11-26 19:21:10.596372] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.099 [2024-11-26 19:21:10.603872] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.099 [2024-11-26 19:21:10.603881] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.099 [2024-11-26 19:21:10.603885] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.603890] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.099 [2024-11-26 19:21:10.603901] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:24:58.099 [2024-11-26 19:21:10.603908] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read vs (no timeout) 00:24:58.099 [2024-11-26 19:21:10.603917] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read vs wait for vs (no timeout) 00:24:58.099 [2024-11-26 19:21:10.603931] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.603935] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.603939] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.099 [2024-11-26 19:21:10.603947] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.099 [2024-11-26 19:21:10.603960] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.099 [2024-11-26 19:21:10.604121] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.099 [2024-11-26 19:21:10.604128] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.099 [2024-11-26 19:21:10.604131] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604135] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.099 [2024-11-26 19:21:10.604142] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read cap (no timeout) 00:24:58.099 [2024-11-26 19:21:10.604150] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read cap wait for cap (no timeout) 00:24:58.099 [2024-11-26 19:21:10.604157] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604161] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604164] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.099 [2024-11-26 19:21:10.604171] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.099 [2024-11-26 19:21:10.604182] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.099 [2024-11-26 19:21:10.604368] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.099 [2024-11-26 19:21:10.604374] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.099 [2024-11-26 19:21:10.604378] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604382] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.099 [2024-11-26 19:21:10.604387] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to check en (no timeout) 00:24:58.099 [2024-11-26 19:21:10.604395] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to check en wait for cc (timeout 15000 ms) 00:24:58.099 [2024-11-26 19:21:10.604402] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604406] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604409] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.099 [2024-11-26 19:21:10.604416] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.099 [2024-11-26 19:21:10.604426] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.099 [2024-11-26 19:21:10.604628] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.099 [2024-11-26 19:21:10.604634] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.099 [2024-11-26 19:21:10.604638] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604642] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.099 [2024-11-26 19:21:10.604646] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:24:58.099 [2024-11-26 19:21:10.604656] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604662] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604666] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.099 [2024-11-26 19:21:10.604673] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.099 [2024-11-26 19:21:10.604683] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.099 [2024-11-26 19:21:10.604888] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.099 [2024-11-26 19:21:10.604895] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.099 [2024-11-26 19:21:10.604899] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.604903] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.099 [2024-11-26 19:21:10.604907] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CC.EN = 0 && CSTS.RDY = 0 00:24:58.099 [2024-11-26 19:21:10.604912] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to controller is disabled (timeout 15000 ms) 00:24:58.099 [2024-11-26 19:21:10.604920] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:24:58.099 [2024-11-26 19:21:10.605028] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Setting CC.EN = 1 00:24:58.099 [2024-11-26 19:21:10.605033] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:24:58.099 [2024-11-26 19:21:10.605040] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.605044] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.605048] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.099 [2024-11-26 19:21:10.605055] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.099 [2024-11-26 19:21:10.605065] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.099 [2024-11-26 19:21:10.605247] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.099 [2024-11-26 19:21:10.605253] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.099 [2024-11-26 19:21:10.605257] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.099 [2024-11-26 19:21:10.605261] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.099 [2024-11-26 19:21:10.605265] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:24:58.100 [2024-11-26 19:21:10.605275] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.605279] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.605282] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.605289] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.100 [2024-11-26 19:21:10.605299] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.100 [2024-11-26 19:21:10.605483] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.100 [2024-11-26 19:21:10.605489] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.100 [2024-11-26 19:21:10.605493] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.605497] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.100 [2024-11-26 19:21:10.605501] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:24:58.100 [2024-11-26 19:21:10.605508] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to reset admin queue (timeout 30000 ms) 00:24:58.100 [2024-11-26 19:21:10.605516] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify controller (no timeout) 00:24:58.100 [2024-11-26 19:21:10.605525] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify controller (timeout 30000 ms) 00:24:58.100 [2024-11-26 19:21:10.605533] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.605537] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.605544] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.100 [2024-11-26 19:21:10.605555] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.100 [2024-11-26 19:21:10.605770] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.100 [2024-11-26 19:21:10.605777] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.100 [2024-11-26 19:21:10.605781] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.605785] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=4096, cccid=0 00:24:58.100 [2024-11-26 19:21:10.605789] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691100) on tqpair(0x162f550): expected_datao=0, payload_size=4096 00:24:58.100 [2024-11-26 19:21:10.605794] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.605809] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.605813] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646024] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.100 [2024-11-26 19:21:10.646036] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.100 [2024-11-26 19:21:10.646040] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646044] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.100 [2024-11-26 19:21:10.646052] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] transport max_xfer_size 4294967295 00:24:58.100 [2024-11-26 19:21:10.646057] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] MDTS max_xfer_size 131072 00:24:58.100 [2024-11-26 19:21:10.646062] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CNTLID 0x0001 00:24:58.100 [2024-11-26 19:21:10.646066] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] transport max_sges 16 00:24:58.100 [2024-11-26 19:21:10.646071] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] fuses compare and write: 1 00:24:58.100 [2024-11-26 19:21:10.646076] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to configure AER (timeout 30000 ms) 00:24:58.100 [2024-11-26 19:21:10.646084] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for configure aer (timeout 30000 ms) 00:24:58.100 [2024-11-26 19:21:10.646091] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646095] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646099] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.646106] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:24:58.100 [2024-11-26 19:21:10.646119] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.100 [2024-11-26 19:21:10.646319] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.100 [2024-11-26 19:21:10.646329] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.100 [2024-11-26 19:21:10.646332] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646336] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.100 [2024-11-26 19:21:10.646343] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646347] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646350] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.646356] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:58.100 [2024-11-26 19:21:10.646363] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646367] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646370] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.646376] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:58.100 [2024-11-26 19:21:10.646382] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646386] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646390] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.646396] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:58.100 [2024-11-26 19:21:10.646402] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646405] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646409] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.646415] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:58.100 [2024-11-26 19:21:10.646419] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set keep alive timeout (timeout 30000 ms) 00:24:58.100 [2024-11-26 19:21:10.646430] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:24:58.100 [2024-11-26 19:21:10.646437] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646440] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x162f550) 00:24:58.100 [2024-11-26 19:21:10.646447] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.100 [2024-11-26 19:21:10.646459] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691100, cid 0, qid 0 00:24:58.100 [2024-11-26 19:21:10.646465] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691280, cid 1, qid 0 00:24:58.100 [2024-11-26 19:21:10.646469] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691400, cid 2, qid 0 00:24:58.100 [2024-11-26 19:21:10.646474] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.100 [2024-11-26 19:21:10.646479] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691700, cid 4, qid 0 00:24:58.100 [2024-11-26 19:21:10.646676] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.100 [2024-11-26 19:21:10.646682] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.100 [2024-11-26 19:21:10.646685] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.100 [2024-11-26 19:21:10.646689] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691700) on tqpair=0x162f550 00:24:58.100 [2024-11-26 19:21:10.646694] nvme_ctrlr.c:3059:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Sending keep alive every 5000000 us 00:24:58.100 [2024-11-26 19:21:10.646701] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify controller iocs specific (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.646711] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set number of queues (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.646717] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for set number of queues (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.646724] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.646728] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.646731] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x162f550) 00:24:58.101 [2024-11-26 19:21:10.646738] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:4 cdw10:00000007 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:24:58.101 [2024-11-26 19:21:10.646748] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691700, cid 4, qid 0 00:24:58.101 [2024-11-26 19:21:10.650872] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.101 [2024-11-26 19:21:10.650880] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.101 [2024-11-26 19:21:10.650883] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.650887] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691700) on tqpair=0x162f550 00:24:58.101 [2024-11-26 19:21:10.650953] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify active ns (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.650963] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify active ns (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.650971] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.650975] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x162f550) 00:24:58.101 [2024-11-26 19:21:10.650982] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.101 [2024-11-26 19:21:10.650994] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691700, cid 4, qid 0 00:24:58.101 [2024-11-26 19:21:10.651157] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.101 [2024-11-26 19:21:10.651163] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.101 [2024-11-26 19:21:10.651167] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651170] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=4096, cccid=4 00:24:58.101 [2024-11-26 19:21:10.651175] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691700) on tqpair(0x162f550): expected_datao=0, payload_size=4096 00:24:58.101 [2024-11-26 19:21:10.651179] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651199] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651203] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651358] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.101 [2024-11-26 19:21:10.651364] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.101 [2024-11-26 19:21:10.651368] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651371] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691700) on tqpair=0x162f550 00:24:58.101 [2024-11-26 19:21:10.651383] nvme_ctrlr.c:4735:spdk_nvme_ctrlr_get_ns: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Namespace 1 was added 00:24:58.101 [2024-11-26 19:21:10.651396] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify ns (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.651406] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify ns (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.651414] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651418] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x162f550) 00:24:58.101 [2024-11-26 19:21:10.651425] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000000 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.101 [2024-11-26 19:21:10.651436] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691700, cid 4, qid 0 00:24:58.101 [2024-11-26 19:21:10.651670] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.101 [2024-11-26 19:21:10.651677] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.101 [2024-11-26 19:21:10.651680] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651684] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=4096, cccid=4 00:24:58.101 [2024-11-26 19:21:10.651688] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691700) on tqpair(0x162f550): expected_datao=0, payload_size=4096 00:24:58.101 [2024-11-26 19:21:10.651692] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651699] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651703] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651894] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.101 [2024-11-26 19:21:10.651901] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.101 [2024-11-26 19:21:10.651904] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651908] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691700) on tqpair=0x162f550 00:24:58.101 [2024-11-26 19:21:10.651918] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.651927] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.651934] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.651938] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x162f550) 00:24:58.101 [2024-11-26 19:21:10.651944] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.101 [2024-11-26 19:21:10.651955] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691700, cid 4, qid 0 00:24:58.101 [2024-11-26 19:21:10.652151] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.101 [2024-11-26 19:21:10.652157] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.101 [2024-11-26 19:21:10.652161] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.652164] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=4096, cccid=4 00:24:58.101 [2024-11-26 19:21:10.652169] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691700) on tqpair(0x162f550): expected_datao=0, payload_size=4096 00:24:58.101 [2024-11-26 19:21:10.652173] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.652220] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.652224] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.652360] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.101 [2024-11-26 19:21:10.652366] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.101 [2024-11-26 19:21:10.652369] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.652373] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691700) on tqpair=0x162f550 00:24:58.101 [2024-11-26 19:21:10.652383] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify ns iocs specific (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.652393] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set supported log pages (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.652400] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set supported features (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.652407] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set host behavior support feature (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.652412] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set doorbell buffer config (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.652417] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set host ID (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.652422] nvme_ctrlr.c:3147:nvme_ctrlr_set_host_id: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] NVMe-oF transport - not sending Set Features - Host ID 00:24:58.101 [2024-11-26 19:21:10.652427] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to transport ready (timeout 30000 ms) 00:24:58.101 [2024-11-26 19:21:10.652432] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to ready (no timeout) 00:24:58.101 [2024-11-26 19:21:10.652446] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.101 [2024-11-26 19:21:10.652450] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x162f550) 00:24:58.101 [2024-11-26 19:21:10.652456] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:4 cdw10:00000001 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.652463] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.652467] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.652471] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.652477] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:24:58.102 [2024-11-26 19:21:10.652489] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691700, cid 4, qid 0 00:24:58.102 [2024-11-26 19:21:10.652495] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691880, cid 5, qid 0 00:24:58.102 [2024-11-26 19:21:10.652695] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.102 [2024-11-26 19:21:10.652701] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.102 [2024-11-26 19:21:10.652704] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.652708] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691700) on tqpair=0x162f550 00:24:58.102 [2024-11-26 19:21:10.652715] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.102 [2024-11-26 19:21:10.652721] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.102 [2024-11-26 19:21:10.652724] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.652728] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691880) on tqpair=0x162f550 00:24:58.102 [2024-11-26 19:21:10.652737] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.652741] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.652747] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:5 cdw10:00000002 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.652757] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691880, cid 5, qid 0 00:24:58.102 [2024-11-26 19:21:10.652932] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.102 [2024-11-26 19:21:10.652939] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.102 [2024-11-26 19:21:10.652944] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.652948] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691880) on tqpair=0x162f550 00:24:58.102 [2024-11-26 19:21:10.652957] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.652961] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.652967] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:5 cdw10:00000004 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.652978] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691880, cid 5, qid 0 00:24:58.102 [2024-11-26 19:21:10.653201] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.102 [2024-11-26 19:21:10.653207] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.102 [2024-11-26 19:21:10.653211] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653215] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691880) on tqpair=0x162f550 00:24:58.102 [2024-11-26 19:21:10.653224] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653227] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.653234] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:5 cdw10:00000007 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.653244] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691880, cid 5, qid 0 00:24:58.102 [2024-11-26 19:21:10.653426] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.102 [2024-11-26 19:21:10.653432] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.102 [2024-11-26 19:21:10.653436] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653440] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691880) on tqpair=0x162f550 00:24:58.102 [2024-11-26 19:21:10.653453] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653457] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.653464] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:5 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.653471] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653475] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.653481] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:ffffffff cdw10:007f0002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.653488] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653492] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=6 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.653498] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:6 nsid:ffffffff cdw10:007f0003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.653506] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653509] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x162f550) 00:24:58.102 [2024-11-26 19:21:10.653516] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:7 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.102 [2024-11-26 19:21:10.653527] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691880, cid 5, qid 0 00:24:58.102 [2024-11-26 19:21:10.653532] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691700, cid 4, qid 0 00:24:58.102 [2024-11-26 19:21:10.653537] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691a00, cid 6, qid 0 00:24:58.102 [2024-11-26 19:21:10.653545] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691b80, cid 7, qid 0 00:24:58.102 [2024-11-26 19:21:10.653799] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.102 [2024-11-26 19:21:10.653805] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.102 [2024-11-26 19:21:10.653809] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653812] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=8192, cccid=5 00:24:58.102 [2024-11-26 19:21:10.653817] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691880) on tqpair(0x162f550): expected_datao=0, payload_size=8192 00:24:58.102 [2024-11-26 19:21:10.653821] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653892] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653897] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653903] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.102 [2024-11-26 19:21:10.653908] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.102 [2024-11-26 19:21:10.653912] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.102 [2024-11-26 19:21:10.653915] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=512, cccid=4 00:24:58.103 [2024-11-26 19:21:10.653920] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691700) on tqpair(0x162f550): expected_datao=0, payload_size=512 00:24:58.103 [2024-11-26 19:21:10.653924] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653930] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653934] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653940] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.103 [2024-11-26 19:21:10.653945] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.103 [2024-11-26 19:21:10.653949] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653952] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=512, cccid=6 00:24:58.103 [2024-11-26 19:21:10.653956] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691a00) on tqpair(0x162f550): expected_datao=0, payload_size=512 00:24:58.103 [2024-11-26 19:21:10.653961] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653967] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653970] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653976] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:24:58.103 [2024-11-26 19:21:10.653982] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:24:58.103 [2024-11-26 19:21:10.653985] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.653989] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x162f550): datao=0, datal=4096, cccid=7 00:24:58.103 [2024-11-26 19:21:10.653993] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1691b80) on tqpair(0x162f550): expected_datao=0, payload_size=4096 00:24:58.103 [2024-11-26 19:21:10.653997] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.654009] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.654013] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.654194] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.103 [2024-11-26 19:21:10.654200] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.103 [2024-11-26 19:21:10.654203] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.654207] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691880) on tqpair=0x162f550 00:24:58.103 [2024-11-26 19:21:10.654219] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.103 [2024-11-26 19:21:10.654226] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.103 [2024-11-26 19:21:10.654230] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.654233] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691700) on tqpair=0x162f550 00:24:58.103 [2024-11-26 19:21:10.654243] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.103 [2024-11-26 19:21:10.654249] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.103 [2024-11-26 19:21:10.654252] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.654256] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691a00) on tqpair=0x162f550 00:24:58.103 [2024-11-26 19:21:10.654263] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.103 [2024-11-26 19:21:10.654269] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.103 [2024-11-26 19:21:10.654273] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.103 [2024-11-26 19:21:10.654276] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691b80) on tqpair=0x162f550 00:24:58.103 ===================================================== 00:24:58.103 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:24:58.103 ===================================================== 00:24:58.103 Controller Capabilities/Features 00:24:58.103 ================================ 00:24:58.103 Vendor ID: 8086 00:24:58.103 Subsystem Vendor ID: 8086 00:24:58.103 Serial Number: SPDK00000000000001 00:24:58.103 Model Number: SPDK bdev Controller 00:24:58.103 Firmware Version: 25.01 00:24:58.103 Recommended Arb Burst: 6 00:24:58.103 IEEE OUI Identifier: e4 d2 5c 00:24:58.103 Multi-path I/O 00:24:58.103 May have multiple subsystem ports: Yes 00:24:58.103 May have multiple controllers: Yes 00:24:58.103 Associated with SR-IOV VF: No 00:24:58.103 Max Data Transfer Size: 131072 00:24:58.103 Max Number of Namespaces: 32 00:24:58.103 Max Number of I/O Queues: 127 00:24:58.103 NVMe Specification Version (VS): 1.3 00:24:58.103 NVMe Specification Version (Identify): 1.3 00:24:58.103 Maximum Queue Entries: 128 00:24:58.103 Contiguous Queues Required: Yes 00:24:58.103 Arbitration Mechanisms Supported 00:24:58.103 Weighted Round Robin: Not Supported 00:24:58.103 Vendor Specific: Not Supported 00:24:58.103 Reset Timeout: 15000 ms 00:24:58.103 Doorbell Stride: 4 bytes 00:24:58.103 NVM Subsystem Reset: Not Supported 00:24:58.103 Command Sets Supported 00:24:58.103 NVM Command Set: Supported 00:24:58.103 Boot Partition: Not Supported 00:24:58.103 Memory Page Size Minimum: 4096 bytes 00:24:58.103 Memory Page Size Maximum: 4096 bytes 00:24:58.103 Persistent Memory Region: Not Supported 00:24:58.103 Optional Asynchronous Events Supported 00:24:58.103 Namespace Attribute Notices: Supported 00:24:58.103 Firmware Activation Notices: Not Supported 00:24:58.103 ANA Change Notices: Not Supported 00:24:58.103 PLE Aggregate Log Change Notices: Not Supported 00:24:58.103 LBA Status Info Alert Notices: Not Supported 00:24:58.103 EGE Aggregate Log Change Notices: Not Supported 00:24:58.103 Normal NVM Subsystem Shutdown event: Not Supported 00:24:58.103 Zone Descriptor Change Notices: Not Supported 00:24:58.103 Discovery Log Change Notices: Not Supported 00:24:58.103 Controller Attributes 00:24:58.103 128-bit Host Identifier: Supported 00:24:58.103 Non-Operational Permissive Mode: Not Supported 00:24:58.103 NVM Sets: Not Supported 00:24:58.103 Read Recovery Levels: Not Supported 00:24:58.103 Endurance Groups: Not Supported 00:24:58.103 Predictable Latency Mode: Not Supported 00:24:58.103 Traffic Based Keep ALive: Not Supported 00:24:58.103 Namespace Granularity: Not Supported 00:24:58.103 SQ Associations: Not Supported 00:24:58.103 UUID List: Not Supported 00:24:58.103 Multi-Domain Subsystem: Not Supported 00:24:58.103 Fixed Capacity Management: Not Supported 00:24:58.103 Variable Capacity Management: Not Supported 00:24:58.103 Delete Endurance Group: Not Supported 00:24:58.103 Delete NVM Set: Not Supported 00:24:58.103 Extended LBA Formats Supported: Not Supported 00:24:58.103 Flexible Data Placement Supported: Not Supported 00:24:58.103 00:24:58.103 Controller Memory Buffer Support 00:24:58.103 ================================ 00:24:58.103 Supported: No 00:24:58.103 00:24:58.103 Persistent Memory Region Support 00:24:58.103 ================================ 00:24:58.103 Supported: No 00:24:58.103 00:24:58.103 Admin Command Set Attributes 00:24:58.103 ============================ 00:24:58.103 Security Send/Receive: Not Supported 00:24:58.103 Format NVM: Not Supported 00:24:58.103 Firmware Activate/Download: Not Supported 00:24:58.103 Namespace Management: Not Supported 00:24:58.103 Device Self-Test: Not Supported 00:24:58.103 Directives: Not Supported 00:24:58.103 NVMe-MI: Not Supported 00:24:58.103 Virtualization Management: Not Supported 00:24:58.103 Doorbell Buffer Config: Not Supported 00:24:58.103 Get LBA Status Capability: Not Supported 00:24:58.103 Command & Feature Lockdown Capability: Not Supported 00:24:58.103 Abort Command Limit: 4 00:24:58.103 Async Event Request Limit: 4 00:24:58.103 Number of Firmware Slots: N/A 00:24:58.104 Firmware Slot 1 Read-Only: N/A 00:24:58.104 Firmware Activation Without Reset: N/A 00:24:58.104 Multiple Update Detection Support: N/A 00:24:58.104 Firmware Update Granularity: No Information Provided 00:24:58.104 Per-Namespace SMART Log: No 00:24:58.104 Asymmetric Namespace Access Log Page: Not Supported 00:24:58.104 Subsystem NQN: nqn.2016-06.io.spdk:cnode1 00:24:58.104 Command Effects Log Page: Supported 00:24:58.104 Get Log Page Extended Data: Supported 00:24:58.104 Telemetry Log Pages: Not Supported 00:24:58.104 Persistent Event Log Pages: Not Supported 00:24:58.104 Supported Log Pages Log Page: May Support 00:24:58.104 Commands Supported & Effects Log Page: Not Supported 00:24:58.104 Feature Identifiers & Effects Log Page:May Support 00:24:58.104 NVMe-MI Commands & Effects Log Page: May Support 00:24:58.104 Data Area 4 for Telemetry Log: Not Supported 00:24:58.104 Error Log Page Entries Supported: 128 00:24:58.104 Keep Alive: Supported 00:24:58.104 Keep Alive Granularity: 10000 ms 00:24:58.104 00:24:58.104 NVM Command Set Attributes 00:24:58.104 ========================== 00:24:58.104 Submission Queue Entry Size 00:24:58.104 Max: 64 00:24:58.104 Min: 64 00:24:58.104 Completion Queue Entry Size 00:24:58.104 Max: 16 00:24:58.104 Min: 16 00:24:58.104 Number of Namespaces: 32 00:24:58.104 Compare Command: Supported 00:24:58.104 Write Uncorrectable Command: Not Supported 00:24:58.104 Dataset Management Command: Supported 00:24:58.104 Write Zeroes Command: Supported 00:24:58.104 Set Features Save Field: Not Supported 00:24:58.104 Reservations: Supported 00:24:58.104 Timestamp: Not Supported 00:24:58.104 Copy: Supported 00:24:58.104 Volatile Write Cache: Present 00:24:58.104 Atomic Write Unit (Normal): 1 00:24:58.104 Atomic Write Unit (PFail): 1 00:24:58.104 Atomic Compare & Write Unit: 1 00:24:58.104 Fused Compare & Write: Supported 00:24:58.104 Scatter-Gather List 00:24:58.104 SGL Command Set: Supported 00:24:58.104 SGL Keyed: Supported 00:24:58.104 SGL Bit Bucket Descriptor: Not Supported 00:24:58.104 SGL Metadata Pointer: Not Supported 00:24:58.104 Oversized SGL: Not Supported 00:24:58.104 SGL Metadata Address: Not Supported 00:24:58.104 SGL Offset: Supported 00:24:58.104 Transport SGL Data Block: Not Supported 00:24:58.104 Replay Protected Memory Block: Not Supported 00:24:58.104 00:24:58.104 Firmware Slot Information 00:24:58.104 ========================= 00:24:58.104 Active slot: 1 00:24:58.104 Slot 1 Firmware Revision: 25.01 00:24:58.104 00:24:58.104 00:24:58.104 Commands Supported and Effects 00:24:58.104 ============================== 00:24:58.104 Admin Commands 00:24:58.104 -------------- 00:24:58.104 Get Log Page (02h): Supported 00:24:58.104 Identify (06h): Supported 00:24:58.104 Abort (08h): Supported 00:24:58.104 Set Features (09h): Supported 00:24:58.104 Get Features (0Ah): Supported 00:24:58.104 Asynchronous Event Request (0Ch): Supported 00:24:58.104 Keep Alive (18h): Supported 00:24:58.104 I/O Commands 00:24:58.104 ------------ 00:24:58.104 Flush (00h): Supported LBA-Change 00:24:58.104 Write (01h): Supported LBA-Change 00:24:58.104 Read (02h): Supported 00:24:58.104 Compare (05h): Supported 00:24:58.104 Write Zeroes (08h): Supported LBA-Change 00:24:58.104 Dataset Management (09h): Supported LBA-Change 00:24:58.104 Copy (19h): Supported LBA-Change 00:24:58.104 00:24:58.104 Error Log 00:24:58.104 ========= 00:24:58.104 00:24:58.104 Arbitration 00:24:58.104 =========== 00:24:58.104 Arbitration Burst: 1 00:24:58.104 00:24:58.104 Power Management 00:24:58.104 ================ 00:24:58.104 Number of Power States: 1 00:24:58.104 Current Power State: Power State #0 00:24:58.104 Power State #0: 00:24:58.104 Max Power: 0.00 W 00:24:58.104 Non-Operational State: Operational 00:24:58.104 Entry Latency: Not Reported 00:24:58.104 Exit Latency: Not Reported 00:24:58.104 Relative Read Throughput: 0 00:24:58.104 Relative Read Latency: 0 00:24:58.104 Relative Write Throughput: 0 00:24:58.104 Relative Write Latency: 0 00:24:58.104 Idle Power: Not Reported 00:24:58.104 Active Power: Not Reported 00:24:58.104 Non-Operational Permissive Mode: Not Supported 00:24:58.104 00:24:58.104 Health Information 00:24:58.104 ================== 00:24:58.104 Critical Warnings: 00:24:58.104 Available Spare Space: OK 00:24:58.104 Temperature: OK 00:24:58.104 Device Reliability: OK 00:24:58.104 Read Only: No 00:24:58.104 Volatile Memory Backup: OK 00:24:58.104 Current Temperature: 0 Kelvin (-273 Celsius) 00:24:58.104 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:24:58.104 Available Spare: 0% 00:24:58.104 Available Spare Threshold: 0% 00:24:58.104 Life Percentage Used:[2024-11-26 19:21:10.654370] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.104 [2024-11-26 19:21:10.654376] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x162f550) 00:24:58.104 [2024-11-26 19:21:10.654382] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:7 cdw10:00000005 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.104 [2024-11-26 19:21:10.654394] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691b80, cid 7, qid 0 00:24:58.104 [2024-11-26 19:21:10.654549] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.104 [2024-11-26 19:21:10.654556] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.104 [2024-11-26 19:21:10.654559] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.104 [2024-11-26 19:21:10.654563] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691b80) on tqpair=0x162f550 00:24:58.104 [2024-11-26 19:21:10.654594] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Prepare to destruct SSD 00:24:58.104 [2024-11-26 19:21:10.654604] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691100) on tqpair=0x162f550 00:24:58.104 [2024-11-26 19:21:10.654610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.104 [2024-11-26 19:21:10.654615] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691280) on tqpair=0x162f550 00:24:58.104 [2024-11-26 19:21:10.654620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.104 [2024-11-26 19:21:10.654625] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691400) on tqpair=0x162f550 00:24:58.104 [2024-11-26 19:21:10.654630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.104 [2024-11-26 19:21:10.654635] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.104 [2024-11-26 19:21:10.654639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:58.104 [2024-11-26 19:21:10.654647] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.104 [2024-11-26 19:21:10.654651] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.104 [2024-11-26 19:21:10.654655] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.104 [2024-11-26 19:21:10.654662] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.104 [2024-11-26 19:21:10.654673] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.104 [2024-11-26 19:21:10.654834] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.104 [2024-11-26 19:21:10.654840] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.654847] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.654851] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.654858] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.658868] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.658874] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.658881] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.658896] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.659085] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.659091] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.659095] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659098] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.659103] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] RTD3E = 0 us 00:24:58.105 [2024-11-26 19:21:10.659108] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] shutdown timeout = 10000 ms 00:24:58.105 [2024-11-26 19:21:10.659117] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659121] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659124] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.659131] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.659141] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.659316] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.659322] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.659325] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659329] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.659339] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659343] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659346] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.659353] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.659363] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.659541] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.659547] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.659550] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659554] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.659564] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659568] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659571] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.659578] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.659588] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.659784] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.659790] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.659794] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659797] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.659807] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659811] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.659814] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.659821] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.659831] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.660012] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.660019] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.660022] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660026] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.660036] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660039] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660043] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.660050] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.660060] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.660287] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.660294] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.660297] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660301] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.660310] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660314] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660318] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.660324] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.660334] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.660553] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.660559] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.660563] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660566] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.660576] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660580] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660583] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.660590] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.660600] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.660791] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.660797] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.660802] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660806] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.660816] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660820] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.660824] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.105 [2024-11-26 19:21:10.660830] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.105 [2024-11-26 19:21:10.660840] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.105 [2024-11-26 19:21:10.661063] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.105 [2024-11-26 19:21:10.661070] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.105 [2024-11-26 19:21:10.661073] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.661077] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.105 [2024-11-26 19:21:10.661087] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.105 [2024-11-26 19:21:10.661091] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661094] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.661101] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.661111] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.661291] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.661298] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.661301] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661305] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.661314] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661318] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661322] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.661328] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.661338] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.661512] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.661518] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.661522] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661526] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.661535] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661539] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661543] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.661549] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.661559] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.661784] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.661790] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.661793] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661799] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.661809] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661813] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.661816] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.661823] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.661833] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.662047] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.662053] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.662057] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662061] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.662070] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662074] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662078] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.662084] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.662094] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.662309] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.662315] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.662319] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662323] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.662332] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662336] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662339] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.662346] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.662356] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.662551] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.662557] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.662561] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662565] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.662574] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662578] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662582] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.662588] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.662598] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.662794] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.662800] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.662804] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662807] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.662819] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662823] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.662826] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x162f550) 00:24:58.106 [2024-11-26 19:21:10.662833] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:58.106 [2024-11-26 19:21:10.662843] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1691580, cid 3, qid 0 00:24:58.106 [2024-11-26 19:21:10.666871] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:24:58.106 [2024-11-26 19:21:10.666879] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:24:58.106 [2024-11-26 19:21:10.666882] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:24:58.106 [2024-11-26 19:21:10.666886] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1691580) on tqpair=0x162f550 00:24:58.106 [2024-11-26 19:21:10.666894] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] shutdown complete in 7 milliseconds 00:24:58.106 0% 00:24:58.106 Data Units Read: 0 00:24:58.106 Data Units Written: 0 00:24:58.106 Host Read Commands: 0 00:24:58.106 Host Write Commands: 0 00:24:58.106 Controller Busy Time: 0 minutes 00:24:58.106 Power Cycles: 0 00:24:58.106 Power On Hours: 0 hours 00:24:58.106 Unsafe Shutdowns: 0 00:24:58.106 Unrecoverable Media Errors: 0 00:24:58.106 Lifetime Error Log Entries: 0 00:24:58.106 Warning Temperature Time: 0 minutes 00:24:58.106 Critical Temperature Time: 0 minutes 00:24:58.106 00:24:58.106 Number of Queues 00:24:58.106 ================ 00:24:58.106 Number of I/O Submission Queues: 127 00:24:58.106 Number of I/O Completion Queues: 127 00:24:58.106 00:24:58.106 Active Namespaces 00:24:58.106 ================= 00:24:58.106 Namespace ID:1 00:24:58.106 Error Recovery Timeout: Unlimited 00:24:58.106 Command Set Identifier: NVM (00h) 00:24:58.106 Deallocate: Supported 00:24:58.106 Deallocated/Unwritten Error: Not Supported 00:24:58.106 Deallocated Read Value: Unknown 00:24:58.106 Deallocate in Write Zeroes: Not Supported 00:24:58.106 Deallocated Guard Field: 0xFFFF 00:24:58.106 Flush: Supported 00:24:58.106 Reservation: Supported 00:24:58.106 Namespace Sharing Capabilities: Multiple Controllers 00:24:58.106 Size (in LBAs): 131072 (0GiB) 00:24:58.106 Capacity (in LBAs): 131072 (0GiB) 00:24:58.106 Utilization (in LBAs): 131072 (0GiB) 00:24:58.106 NGUID: ABCDEF0123456789ABCDEF0123456789 00:24:58.107 EUI64: ABCDEF0123456789 00:24:58.107 UUID: 94e86b4e-3725-47e6-ad26-4d43833a09c9 00:24:58.107 Thin Provisioning: Not Supported 00:24:58.107 Per-NS Atomic Units: Yes 00:24:58.107 Atomic Boundary Size (Normal): 0 00:24:58.107 Atomic Boundary Size (PFail): 0 00:24:58.107 Atomic Boundary Offset: 0 00:24:58.107 Maximum Single Source Range Length: 65535 00:24:58.107 Maximum Copy Length: 65535 00:24:58.107 Maximum Source Range Count: 1 00:24:58.107 NGUID/EUI64 Never Reused: No 00:24:58.107 Namespace Write Protected: No 00:24:58.107 Number of LBA Formats: 1 00:24:58.107 Current LBA Format: LBA Format #00 00:24:58.107 LBA Format #00: Data Size: 512 Metadata Size: 0 00:24:58.107 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@51 -- # sync 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@52 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@54 -- # trap - SIGINT SIGTERM EXIT 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@56 -- # nvmftestfini 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@516 -- # nvmfcleanup 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@121 -- # sync 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@124 -- # set +e 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:58.107 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:58.107 rmmod nvme_tcp 00:24:58.368 rmmod nvme_fabrics 00:24:58.368 rmmod nvme_keyring 00:24:58.368 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:58.368 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@128 -- # set -e 00:24:58.368 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@129 -- # return 0 00:24:58.368 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@517 -- # '[' -n 3961354 ']' 00:24:58.368 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@518 -- # killprocess 3961354 00:24:58.368 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@954 -- # '[' -z 3961354 ']' 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@958 -- # kill -0 3961354 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@959 -- # uname 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3961354 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3961354' 00:24:58.369 killing process with pid 3961354 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@973 -- # kill 3961354 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@978 -- # wait 3961354 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@297 -- # iptr 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@791 -- # iptables-save 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@791 -- # iptables-restore 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:58.369 19:21:10 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:00.917 00:25:00.917 real 0m12.291s 00:25:00.917 user 0m8.565s 00:25:00.917 sys 0m6.623s 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:25:00.917 ************************************ 00:25:00.917 END TEST nvmf_identify 00:25:00.917 ************************************ 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@23 -- # run_test nvmf_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:25:00.917 ************************************ 00:25:00.917 START TEST nvmf_perf 00:25:00.917 ************************************ 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:25:00.917 * Looking for test storage... 00:25:00.917 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1693 -- # lcov --version 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # IFS=.-: 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # read -ra ver1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # IFS=.-: 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # read -ra ver2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@338 -- # local 'op=<' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@340 -- # ver1_l=2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@341 -- # ver2_l=1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@344 -- # case "$op" in 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@345 -- # : 1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # decimal 1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # ver1[v]=1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # decimal 2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # ver2[v]=2 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # return 0 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:00.917 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:00.917 --rc genhtml_branch_coverage=1 00:25:00.917 --rc genhtml_function_coverage=1 00:25:00.917 --rc genhtml_legend=1 00:25:00.917 --rc geninfo_all_blocks=1 00:25:00.917 --rc geninfo_unexecuted_blocks=1 00:25:00.917 00:25:00.917 ' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:00.917 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:00.917 --rc genhtml_branch_coverage=1 00:25:00.917 --rc genhtml_function_coverage=1 00:25:00.917 --rc genhtml_legend=1 00:25:00.917 --rc geninfo_all_blocks=1 00:25:00.917 --rc geninfo_unexecuted_blocks=1 00:25:00.917 00:25:00.917 ' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:00.917 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:00.917 --rc genhtml_branch_coverage=1 00:25:00.917 --rc genhtml_function_coverage=1 00:25:00.917 --rc genhtml_legend=1 00:25:00.917 --rc geninfo_all_blocks=1 00:25:00.917 --rc geninfo_unexecuted_blocks=1 00:25:00.917 00:25:00.917 ' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:00.917 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:00.917 --rc genhtml_branch_coverage=1 00:25:00.917 --rc genhtml_function_coverage=1 00:25:00.917 --rc genhtml_legend=1 00:25:00.917 --rc geninfo_all_blocks=1 00:25:00.917 --rc geninfo_unexecuted_blocks=1 00:25:00.917 00:25:00.917 ' 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # uname -s 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:00.917 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@15 -- # shopt -s extglob 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@5 -- # export PATH 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@51 -- # : 0 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:00.918 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@12 -- # MALLOC_BDEV_SIZE=64 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@17 -- # nvmftestinit 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@476 -- # prepare_net_devs 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@438 -- # local -g is_hw=no 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # remove_spdk_ns 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@309 -- # xtrace_disable 00:25:00.918 19:21:13 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # pci_devs=() 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # net_devs=() 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # e810=() 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # local -ga e810 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # x722=() 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # local -ga x722 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # mlx=() 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # local -ga mlx 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:25:09.054 Found 0000:31:00.0 (0x8086 - 0x159b) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:25:09.054 Found 0000:31:00.1 (0x8086 - 0x159b) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:09.054 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:25:09.055 Found net devices under 0000:31:00.0: cvl_0_0 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:25:09.055 Found net devices under 0000:31:00.1: cvl_0_1 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # is_hw=yes 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:09.055 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:09.315 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:09.315 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:09.315 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:09.315 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:09.315 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:09.315 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:09.315 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:09.575 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:09.575 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.492 ms 00:25:09.575 00:25:09.575 --- 10.0.0.2 ping statistics --- 00:25:09.575 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:09.575 rtt min/avg/max/mdev = 0.492/0.492/0.492/0.000 ms 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:09.575 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:09.575 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.290 ms 00:25:09.575 00:25:09.575 --- 10.0.0.1 ping statistics --- 00:25:09.575 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:09.575 rtt min/avg/max/mdev = 0.290/0.290/0.290/0.000 ms 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@450 -- # return 0 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@18 -- # nvmfappstart -m 0xF 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:25:09.575 19:21:21 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@726 -- # xtrace_disable 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@509 -- # nvmfpid=3966669 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@510 -- # waitforlisten 3966669 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@835 -- # '[' -z 3966669 ']' 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:09.575 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:09.575 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:25:09.575 [2024-11-26 19:21:22.066912] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:25:09.575 [2024-11-26 19:21:22.066979] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:09.575 [2024-11-26 19:21:22.157383] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:09.835 [2024-11-26 19:21:22.198943] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:09.835 [2024-11-26 19:21:22.198985] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:09.835 [2024-11-26 19:21:22.198993] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:09.835 [2024-11-26 19:21:22.199000] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:09.835 [2024-11-26 19:21:22.199005] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:09.835 [2024-11-26 19:21:22.200632] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:09.835 [2024-11-26 19:21:22.200748] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:09.835 [2024-11-26 19:21:22.200917] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:09.835 [2024-11-26 19:21:22.200917] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@868 -- # return 0 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@732 -- # xtrace_disable 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:25:10.404 19:21:22 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:25:10.973 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_get_config bdev 00:25:10.973 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # jq -r '.[].params | select(.name=="Nvme0").traddr' 00:25:11.233 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # local_nvme_trid=0000:65:00.0 00:25:11.233 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:25:11.233 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # bdevs=' Malloc0' 00:25:11.233 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@33 -- # '[' -n 0000:65:00.0 ']' 00:25:11.233 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@34 -- # bdevs=' Malloc0 Nvme0n1' 00:25:11.233 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@37 -- # '[' tcp == rdma ']' 00:25:11.233 19:21:23 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:25:11.493 [2024-11-26 19:21:23.969991] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:11.493 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:11.752 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:25:11.752 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:11.752 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:25:11.752 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:25:12.012 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:12.271 [2024-11-26 19:21:24.712742] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:12.271 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:25:12.620 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@52 -- # '[' -n 0000:65:00.0 ']' 00:25:12.620 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@53 -- # perf_app -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:65:00.0' 00:25:12.620 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@21 -- # '[' 0 -eq 1 ']' 00:25:12.620 19:21:24 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:65:00.0' 00:25:13.611 Initializing NVMe Controllers 00:25:13.611 Attached to NVMe Controller at 0000:65:00.0 [144d:a80a] 00:25:13.612 Associating PCIE (0000:65:00.0) NSID 1 with lcore 0 00:25:13.612 Initialization complete. Launching workers. 00:25:13.612 ======================================================== 00:25:13.612 Latency(us) 00:25:13.612 Device Information : IOPS MiB/s Average min max 00:25:13.612 PCIE (0000:65:00.0) NSID 1 from core 0: 79050.48 308.79 404.07 13.36 4889.92 00:25:13.612 ======================================================== 00:25:13.612 Total : 79050.48 308.79 404.07 13.36 4889.92 00:25:13.612 00:25:13.612 19:21:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:25:14.990 Initializing NVMe Controllers 00:25:14.990 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:14.990 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:25:14.990 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:25:14.990 Initialization complete. Launching workers. 00:25:14.990 ======================================================== 00:25:14.990 Latency(us) 00:25:14.990 Device Information : IOPS MiB/s Average min max 00:25:14.990 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 96.00 0.37 10748.35 260.85 46220.20 00:25:14.991 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 49.00 0.19 21011.21 7964.95 47903.47 00:25:14.991 ======================================================== 00:25:14.991 Total : 145.00 0.57 14216.49 260.85 47903.47 00:25:14.991 00:25:14.991 19:21:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 4096 -w randrw -M 50 -t 1 -HI -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:25:16.370 Initializing NVMe Controllers 00:25:16.370 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:16.370 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:25:16.370 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:25:16.370 Initialization complete. Launching workers. 00:25:16.370 ======================================================== 00:25:16.371 Latency(us) 00:25:16.371 Device Information : IOPS MiB/s Average min max 00:25:16.371 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 10470.00 40.90 3059.33 495.53 6560.68 00:25:16.371 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 3871.00 15.12 8327.09 6072.24 15760.76 00:25:16.371 ======================================================== 00:25:16.371 Total : 14341.00 56.02 4481.23 495.53 15760.76 00:25:16.371 00:25:16.371 19:21:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ e810 == \e\8\1\0 ]] 00:25:16.371 19:21:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ tcp == \r\d\m\a ]] 00:25:16.371 19:21:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -O 16384 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:25:18.915 Initializing NVMe Controllers 00:25:18.915 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:18.915 Controller IO queue size 128, less than required. 00:25:18.915 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:25:18.915 Controller IO queue size 128, less than required. 00:25:18.915 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:25:18.915 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:25:18.915 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:25:18.915 Initialization complete. Launching workers. 00:25:18.915 ======================================================== 00:25:18.915 Latency(us) 00:25:18.915 Device Information : IOPS MiB/s Average min max 00:25:18.915 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1645.82 411.46 79423.08 53081.47 104152.21 00:25:18.915 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 576.24 144.06 225736.16 88262.74 336771.59 00:25:18.915 ======================================================== 00:25:18.915 Total : 2222.06 555.52 117365.84 53081.47 336771.59 00:25:18.915 00:25:18.915 19:21:31 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 36964 -O 4096 -w randrw -M 50 -t 5 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0xf -P 4 00:25:19.175 No valid NVMe controllers or AIO or URING devices found 00:25:19.175 Initializing NVMe Controllers 00:25:19.175 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:19.175 Controller IO queue size 128, less than required. 00:25:19.175 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:25:19.175 WARNING: IO size 36964 (-o) is not a multiple of nsid 1 sector size 512. Removing this ns from test 00:25:19.175 Controller IO queue size 128, less than required. 00:25:19.175 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:25:19.175 WARNING: IO size 36964 (-o) is not a multiple of nsid 2 sector size 512. Removing this ns from test 00:25:19.175 WARNING: Some requested NVMe devices were skipped 00:25:19.175 19:21:31 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' --transport-stat 00:25:21.716 Initializing NVMe Controllers 00:25:21.716 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:21.716 Controller IO queue size 128, less than required. 00:25:21.716 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:25:21.716 Controller IO queue size 128, less than required. 00:25:21.716 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:25:21.716 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:25:21.716 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:25:21.716 Initialization complete. Launching workers. 00:25:21.716 00:25:21.716 ==================== 00:25:21.716 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 statistics: 00:25:21.716 TCP transport: 00:25:21.716 polls: 22667 00:25:21.716 idle_polls: 13442 00:25:21.716 sock_completions: 9225 00:25:21.716 nvme_completions: 6417 00:25:21.716 submitted_requests: 9648 00:25:21.716 queued_requests: 1 00:25:21.716 00:25:21.716 ==================== 00:25:21.716 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 statistics: 00:25:21.716 TCP transport: 00:25:21.716 polls: 19915 00:25:21.716 idle_polls: 10260 00:25:21.716 sock_completions: 9655 00:25:21.716 nvme_completions: 6869 00:25:21.716 submitted_requests: 10262 00:25:21.716 queued_requests: 1 00:25:21.716 ======================================================== 00:25:21.716 Latency(us) 00:25:21.716 Device Information : IOPS MiB/s Average min max 00:25:21.716 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1603.58 400.89 81188.47 46641.42 139200.23 00:25:21.716 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 1716.55 429.14 75299.63 27933.80 119887.02 00:25:21.716 ======================================================== 00:25:21.716 Total : 3320.12 830.03 78143.86 27933.80 139200.23 00:25:21.716 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@66 -- # sync 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@67 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@69 -- # '[' 0 -eq 1 ']' 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@112 -- # trap - SIGINT SIGTERM EXIT 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@114 -- # nvmftestfini 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@516 -- # nvmfcleanup 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@121 -- # sync 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@124 -- # set +e 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:21.716 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:21.716 rmmod nvme_tcp 00:25:21.716 rmmod nvme_fabrics 00:25:21.995 rmmod nvme_keyring 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@128 -- # set -e 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@129 -- # return 0 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@517 -- # '[' -n 3966669 ']' 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@518 -- # killprocess 3966669 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@954 -- # '[' -z 3966669 ']' 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@958 -- # kill -0 3966669 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@959 -- # uname 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3966669 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3966669' 00:25:21.995 killing process with pid 3966669 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@973 -- # kill 3966669 00:25:21.995 19:21:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@978 -- # wait 3966669 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@297 -- # iptr 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@791 -- # iptables-save 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@791 -- # iptables-restore 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:23.904 19:21:36 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:26.450 00:25:26.450 real 0m25.367s 00:25:26.450 user 0m58.525s 00:25:26.450 sys 0m9.404s 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:25:26.450 ************************************ 00:25:26.450 END TEST nvmf_perf 00:25:26.450 ************************************ 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@24 -- # run_test nvmf_fio_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.450 ************************************ 00:25:26.450 START TEST nvmf_fio_host 00:25:26.450 ************************************ 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:25:26.450 * Looking for test storage... 00:25:26.450 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1693 -- # lcov --version 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # IFS=.-: 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # read -ra ver1 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # IFS=.-: 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # read -ra ver2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@338 -- # local 'op=<' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@340 -- # ver1_l=2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@341 -- # ver2_l=1 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@344 -- # case "$op" in 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@345 -- # : 1 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # decimal 1 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=1 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 1 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # ver1[v]=1 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # decimal 2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # ver2[v]=2 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # return 0 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:26.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:26.450 --rc genhtml_branch_coverage=1 00:25:26.450 --rc genhtml_function_coverage=1 00:25:26.450 --rc genhtml_legend=1 00:25:26.450 --rc geninfo_all_blocks=1 00:25:26.450 --rc geninfo_unexecuted_blocks=1 00:25:26.450 00:25:26.450 ' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:26.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:26.450 --rc genhtml_branch_coverage=1 00:25:26.450 --rc genhtml_function_coverage=1 00:25:26.450 --rc genhtml_legend=1 00:25:26.450 --rc geninfo_all_blocks=1 00:25:26.450 --rc geninfo_unexecuted_blocks=1 00:25:26.450 00:25:26.450 ' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:26.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:26.450 --rc genhtml_branch_coverage=1 00:25:26.450 --rc genhtml_function_coverage=1 00:25:26.450 --rc genhtml_legend=1 00:25:26.450 --rc geninfo_all_blocks=1 00:25:26.450 --rc geninfo_unexecuted_blocks=1 00:25:26.450 00:25:26.450 ' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:26.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:26.450 --rc genhtml_branch_coverage=1 00:25:26.450 --rc genhtml_function_coverage=1 00:25:26.450 --rc genhtml_legend=1 00:25:26.450 --rc geninfo_all_blocks=1 00:25:26.450 --rc geninfo_unexecuted_blocks=1 00:25:26.450 00:25:26.450 ' 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:26.450 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # uname -s 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@51 -- # : 0 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:26.451 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@14 -- # nvmftestinit 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@476 -- # prepare_net_devs 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@438 -- # local -g is_hw=no 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # remove_spdk_ns 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:26.451 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:26.452 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:25:26.452 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:25:26.452 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@309 -- # xtrace_disable 00:25:26.452 19:21:38 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # pci_devs=() 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # net_devs=() 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # e810=() 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # local -ga e810 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # x722=() 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # local -ga x722 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # mlx=() 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # local -ga mlx 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:25:34.586 Found 0000:31:00.0 (0x8086 - 0x159b) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:25:34.586 Found 0000:31:00.1 (0x8086 - 0x159b) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:25:34.586 Found net devices under 0000:31:00.0: cvl_0_0 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:25:34.586 Found net devices under 0000:31:00.1: cvl_0_1 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:34.586 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # is_hw=yes 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:34.587 19:21:46 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:34.587 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:34.587 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:34.587 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:34.587 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:34.848 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:34.848 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.476 ms 00:25:34.848 00:25:34.848 --- 10.0.0.2 ping statistics --- 00:25:34.848 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:34.848 rtt min/avg/max/mdev = 0.476/0.476/0.476/0.000 ms 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:34.848 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:34.848 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.310 ms 00:25:34.848 00:25:34.848 --- 10.0.0.1 ping statistics --- 00:25:34.848 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:34.848 rtt min/avg/max/mdev = 0.310/0.310/0.310/0.000 ms 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@450 -- # return 0 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@16 -- # [[ y != y ]] 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@21 -- # timing_enter start_nvmf_tgt 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@726 -- # xtrace_disable 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@24 -- # nvmfpid=3974349 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@26 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@23 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@28 -- # waitforlisten 3974349 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@835 -- # '[' -z 3974349 ']' 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:34.848 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:34.848 19:21:47 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:25:34.848 [2024-11-26 19:21:47.406521] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:25:34.848 [2024-11-26 19:21:47.406591] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:35.107 [2024-11-26 19:21:47.497090] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:35.107 [2024-11-26 19:21:47.538686] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:35.107 [2024-11-26 19:21:47.538724] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:35.107 [2024-11-26 19:21:47.538732] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:35.107 [2024-11-26 19:21:47.538739] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:35.107 [2024-11-26 19:21:47.538745] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:35.107 [2024-11-26 19:21:47.540369] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:35.107 [2024-11-26 19:21:47.540492] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:35.107 [2024-11-26 19:21:47.540649] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:35.107 [2024-11-26 19:21:47.540649] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:35.676 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:35.676 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@868 -- # return 0 00:25:35.676 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:25:35.936 [2024-11-26 19:21:48.368939] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:35.936 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@30 -- # timing_exit start_nvmf_tgt 00:25:35.936 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@732 -- # xtrace_disable 00:25:35.936 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:25:35.936 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:25:36.195 Malloc1 00:25:36.195 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:36.457 19:21:48 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:25:36.457 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:36.717 [2024-11-26 19:21:49.172645] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:36.717 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:25:36.976 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@38 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:25:36.976 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@41 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:25:36.976 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1364 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:25:36.976 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:25:36.976 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:25:36.976 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local sanitizers 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # shift 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1347 -- # local asan_lib= 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libasan 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:25:36.977 19:21:49 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:25:37.237 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:25:37.237 fio-3.35 00:25:37.237 Starting 1 thread 00:25:39.776 00:25:39.776 test: (groupid=0, jobs=1): err= 0: pid=3974913: Tue Nov 26 19:21:52 2024 00:25:39.776 read: IOPS=13.8k, BW=53.9MiB/s (56.5MB/s)(108MiB/2005msec) 00:25:39.776 slat (usec): min=2, max=279, avg= 2.16, stdev= 2.39 00:25:39.776 clat (usec): min=3321, max=8925, avg=5111.25, stdev=377.47 00:25:39.776 lat (usec): min=3323, max=8927, avg=5113.42, stdev=377.68 00:25:39.776 clat percentiles (usec): 00:25:39.776 | 1.00th=[ 4293], 5.00th=[ 4555], 10.00th=[ 4686], 20.00th=[ 4817], 00:25:39.776 | 30.00th=[ 4948], 40.00th=[ 5014], 50.00th=[ 5080], 60.00th=[ 5211], 00:25:39.776 | 70.00th=[ 5276], 80.00th=[ 5407], 90.00th=[ 5538], 95.00th=[ 5669], 00:25:39.777 | 99.00th=[ 5997], 99.50th=[ 6456], 99.90th=[ 8029], 99.95th=[ 8455], 00:25:39.777 | 99.99th=[ 8717] 00:25:39.777 bw ( KiB/s): min=53876, max=55632, per=99.94%, avg=55131.00, stdev=840.14, samples=4 00:25:39.777 iops : min=13469, max=13908, avg=13782.75, stdev=210.03, samples=4 00:25:39.777 write: IOPS=13.8k, BW=53.8MiB/s (56.4MB/s)(108MiB/2005msec); 0 zone resets 00:25:39.777 slat (usec): min=2, max=270, avg= 2.24, stdev= 1.80 00:25:39.777 clat (usec): min=2677, max=8150, avg=4129.57, stdev=322.60 00:25:39.777 lat (usec): min=2679, max=8152, avg=4131.81, stdev=322.87 00:25:39.777 clat percentiles (usec): 00:25:39.777 | 1.00th=[ 3458], 5.00th=[ 3654], 10.00th=[ 3785], 20.00th=[ 3916], 00:25:39.777 | 30.00th=[ 3982], 40.00th=[ 4047], 50.00th=[ 4113], 60.00th=[ 4178], 00:25:39.777 | 70.00th=[ 4228], 80.00th=[ 4359], 90.00th=[ 4424], 95.00th=[ 4555], 00:25:39.777 | 99.00th=[ 4817], 99.50th=[ 5932], 99.90th=[ 6915], 99.95th=[ 7308], 00:25:39.777 | 99.99th=[ 8029] 00:25:39.777 bw ( KiB/s): min=54235, max=55616, per=100.00%, avg=55094.75, stdev=601.83, samples=4 00:25:39.777 iops : min=13558, max=13904, avg=13773.50, stdev=150.81, samples=4 00:25:39.777 lat (msec) : 4=16.29%, 10=83.71% 00:25:39.777 cpu : usr=74.85%, sys=23.75%, ctx=33, majf=0, minf=16 00:25:39.777 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.9% 00:25:39.777 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:25:39.777 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:25:39.777 issued rwts: total=27650,27617,0,0 short=0,0,0,0 dropped=0,0,0,0 00:25:39.777 latency : target=0, window=0, percentile=100.00%, depth=128 00:25:39.777 00:25:39.777 Run status group 0 (all jobs): 00:25:39.777 READ: bw=53.9MiB/s (56.5MB/s), 53.9MiB/s-53.9MiB/s (56.5MB/s-56.5MB/s), io=108MiB (113MB), run=2005-2005msec 00:25:39.777 WRITE: bw=53.8MiB/s (56.4MB/s), 53.8MiB/s-53.8MiB/s (56.4MB/s-56.4MB/s), io=108MiB (113MB), run=2005-2005msec 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@45 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1364 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local sanitizers 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # shift 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1347 -- # local asan_lib= 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libasan 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:25:39.777 19:21:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:25:40.343 test: (g=0): rw=randrw, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=spdk, iodepth=128 00:25:40.343 fio-3.35 00:25:40.343 Starting 1 thread 00:25:42.886 00:25:42.886 test: (groupid=0, jobs=1): err= 0: pid=3975738: Tue Nov 26 19:21:54 2024 00:25:42.886 read: IOPS=8873, BW=139MiB/s (145MB/s)(278MiB/2006msec) 00:25:42.886 slat (usec): min=3, max=135, avg= 3.94, stdev= 1.93 00:25:42.886 clat (usec): min=1981, max=18226, avg=8805.86, stdev=2239.16 00:25:42.886 lat (usec): min=1986, max=18243, avg=8809.80, stdev=2239.49 00:25:42.886 clat percentiles (usec): 00:25:42.886 | 1.00th=[ 4424], 5.00th=[ 5407], 10.00th=[ 5997], 20.00th=[ 6783], 00:25:42.886 | 30.00th=[ 7504], 40.00th=[ 8160], 50.00th=[ 8717], 60.00th=[ 9372], 00:25:42.886 | 70.00th=[ 9896], 80.00th=[10552], 90.00th=[11731], 95.00th=[12387], 00:25:42.886 | 99.00th=[15139], 99.50th=[16319], 99.90th=[17433], 99.95th=[17695], 00:25:42.886 | 99.99th=[18220] 00:25:42.886 bw ( KiB/s): min=62784, max=78720, per=49.70%, avg=70560.00, stdev=6534.95, samples=4 00:25:42.886 iops : min= 3924, max= 4920, avg=4410.00, stdev=408.43, samples=4 00:25:42.886 write: IOPS=5125, BW=80.1MiB/s (84.0MB/s)(143MiB/1788msec); 0 zone resets 00:25:42.886 slat (usec): min=39, max=580, avg=42.51, stdev=10.45 00:25:42.886 clat (usec): min=1919, max=19583, avg=9985.64, stdev=1736.41 00:25:42.886 lat (usec): min=1959, max=19721, avg=10028.15, stdev=1740.14 00:25:42.886 clat percentiles (usec): 00:25:42.886 | 1.00th=[ 6849], 5.00th=[ 7504], 10.00th=[ 7963], 20.00th=[ 8586], 00:25:42.886 | 30.00th=[ 8979], 40.00th=[ 9503], 50.00th=[ 9896], 60.00th=[10290], 00:25:42.886 | 70.00th=[10683], 80.00th=[11338], 90.00th=[12125], 95.00th=[12780], 00:25:42.886 | 99.00th=[15008], 99.50th=[16450], 99.90th=[19006], 99.95th=[19268], 00:25:42.886 | 99.99th=[19530] 00:25:42.886 bw ( KiB/s): min=64224, max=81280, per=89.40%, avg=73312.00, stdev=7004.76, samples=4 00:25:42.886 iops : min= 4014, max= 5080, avg=4582.00, stdev=437.80, samples=4 00:25:42.886 lat (msec) : 2=0.01%, 4=0.38%, 10=65.11%, 20=34.49% 00:25:42.886 cpu : usr=85.29%, sys=13.36%, ctx=15, majf=0, minf=36 00:25:42.886 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.4%, 32=0.7%, >=64=98.6% 00:25:42.886 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:25:42.886 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:25:42.886 issued rwts: total=17800,9164,0,0 short=0,0,0,0 dropped=0,0,0,0 00:25:42.886 latency : target=0, window=0, percentile=100.00%, depth=128 00:25:42.886 00:25:42.886 Run status group 0 (all jobs): 00:25:42.886 READ: bw=139MiB/s (145MB/s), 139MiB/s-139MiB/s (145MB/s-145MB/s), io=278MiB (292MB), run=2006-2006msec 00:25:42.886 WRITE: bw=80.1MiB/s (84.0MB/s), 80.1MiB/s-80.1MiB/s (84.0MB/s-84.0MB/s), io=143MiB (150MB), run=1788-1788msec 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@49 -- # '[' 0 -eq 1 ']' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@85 -- # rm -f ./local-test-0-verify.state 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@86 -- # nvmftestfini 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@516 -- # nvmfcleanup 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@121 -- # sync 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@124 -- # set +e 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:42.887 rmmod nvme_tcp 00:25:42.887 rmmod nvme_fabrics 00:25:42.887 rmmod nvme_keyring 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@128 -- # set -e 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@129 -- # return 0 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@517 -- # '[' -n 3974349 ']' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@518 -- # killprocess 3974349 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@954 -- # '[' -z 3974349 ']' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@958 -- # kill -0 3974349 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@959 -- # uname 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3974349 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3974349' 00:25:42.887 killing process with pid 3974349 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@973 -- # kill 3974349 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@978 -- # wait 3974349 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@297 -- # iptr 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@791 -- # iptables-save 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@791 -- # iptables-restore 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:42.887 19:21:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:45.431 00:25:45.431 real 0m18.978s 00:25:45.431 user 1m4.990s 00:25:45.431 sys 0m8.337s 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:25:45.431 ************************************ 00:25:45.431 END TEST nvmf_fio_host 00:25:45.431 ************************************ 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@25 -- # run_test nvmf_failover /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:25:45.431 ************************************ 00:25:45.431 START TEST nvmf_failover 00:25:45.431 ************************************ 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:25:45.431 * Looking for test storage... 00:25:45.431 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1693 -- # lcov --version 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # IFS=.-: 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # read -ra ver1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # IFS=.-: 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # read -ra ver2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@338 -- # local 'op=<' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@340 -- # ver1_l=2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@341 -- # ver2_l=1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@344 -- # case "$op" in 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@345 -- # : 1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # decimal 1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # ver1[v]=1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # decimal 2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # ver2[v]=2 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # return 0 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:45.431 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:45.431 --rc genhtml_branch_coverage=1 00:25:45.431 --rc genhtml_function_coverage=1 00:25:45.431 --rc genhtml_legend=1 00:25:45.431 --rc geninfo_all_blocks=1 00:25:45.431 --rc geninfo_unexecuted_blocks=1 00:25:45.431 00:25:45.431 ' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:45.431 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:45.431 --rc genhtml_branch_coverage=1 00:25:45.431 --rc genhtml_function_coverage=1 00:25:45.431 --rc genhtml_legend=1 00:25:45.431 --rc geninfo_all_blocks=1 00:25:45.431 --rc geninfo_unexecuted_blocks=1 00:25:45.431 00:25:45.431 ' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:45.431 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:45.431 --rc genhtml_branch_coverage=1 00:25:45.431 --rc genhtml_function_coverage=1 00:25:45.431 --rc genhtml_legend=1 00:25:45.431 --rc geninfo_all_blocks=1 00:25:45.431 --rc geninfo_unexecuted_blocks=1 00:25:45.431 00:25:45.431 ' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:45.431 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:45.431 --rc genhtml_branch_coverage=1 00:25:45.431 --rc genhtml_function_coverage=1 00:25:45.431 --rc genhtml_legend=1 00:25:45.431 --rc geninfo_all_blocks=1 00:25:45.431 --rc geninfo_unexecuted_blocks=1 00:25:45.431 00:25:45.431 ' 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # uname -s 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:45.431 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@15 -- # shopt -s extglob 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@5 -- # export PATH 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@51 -- # : 0 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:45.432 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@11 -- # MALLOC_BDEV_SIZE=64 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@18 -- # nvmftestinit 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@476 -- # prepare_net_devs 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@438 -- # local -g is_hw=no 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # remove_spdk_ns 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@309 -- # xtrace_disable 00:25:45.432 19:21:57 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # pci_devs=() 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # net_devs=() 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # e810=() 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # local -ga e810 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # x722=() 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # local -ga x722 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # mlx=() 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # local -ga mlx 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:25:53.565 Found 0000:31:00.0 (0x8086 - 0x159b) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:25:53.565 Found 0000:31:00.1 (0x8086 - 0x159b) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:25:53.565 Found net devices under 0000:31:00.0: cvl_0_0 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:25:53.565 Found net devices under 0000:31:00.1: cvl_0_1 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # is_hw=yes 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:53.565 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:53.827 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:53.827 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.623 ms 00:25:53.827 00:25:53.827 --- 10.0.0.2 ping statistics --- 00:25:53.827 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:53.827 rtt min/avg/max/mdev = 0.623/0.623/0.623/0.000 ms 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:53.827 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:53.827 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.320 ms 00:25:53.827 00:25:53.827 --- 10.0.0.1 ping statistics --- 00:25:53.827 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:53.827 rtt min/avg/max/mdev = 0.320/0.320/0.320/0.000 ms 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@450 -- # return 0 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@20 -- # nvmfappstart -m 0xE 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@726 -- # xtrace_disable 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@509 -- # nvmfpid=3980825 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@510 -- # waitforlisten 3980825 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # '[' -z 3980825 ']' 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:53.827 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:53.827 19:22:06 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:25:54.089 [2024-11-26 19:22:06.498452] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:25:54.090 [2024-11-26 19:22:06.498523] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:54.090 [2024-11-26 19:22:06.607561] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:25:54.090 [2024-11-26 19:22:06.658546] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:54.090 [2024-11-26 19:22:06.658601] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:54.090 [2024-11-26 19:22:06.658610] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:54.090 [2024-11-26 19:22:06.658617] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:54.090 [2024-11-26 19:22:06.658623] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:54.090 [2024-11-26 19:22:06.660730] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:54.090 [2024-11-26 19:22:06.660916] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:54.090 [2024-11-26 19:22:06.660917] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@868 -- # return 0 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@732 -- # xtrace_disable 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:25:55.031 [2024-11-26 19:22:07.502152] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:55.031 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:25:55.290 Malloc0 00:25:55.290 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:55.550 19:22:07 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:55.550 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:55.810 [2024-11-26 19:22:08.248454] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:55.810 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:25:55.810 [2024-11-26 19:22:08.432936] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:25:56.070 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:25:56.070 [2024-11-26 19:22:08.617489] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:25:56.070 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@31 -- # bdevperf_pid=3981405 00:25:56.070 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 15 -f 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; cat $testdir/try.txt; rm -f $testdir/try.txt; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@34 -- # waitforlisten 3981405 /var/tmp/bdevperf.sock 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # '[' -z 3981405 ']' 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:25:56.071 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:56.071 19:22:08 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:25:57.009 19:22:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:57.009 19:22:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@868 -- # return 0 00:25:57.009 19:22:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:25:57.269 NVMe0n1 00:25:57.269 19:22:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:25:57.839 00:25:57.839 19:22:10 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@39 -- # run_test_pid=3981653 00:25:57.839 19:22:10 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:25:57.839 19:22:10 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@41 -- # sleep 1 00:25:58.779 19:22:11 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:58.779 [2024-11-26 19:22:11.364732] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364772] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364777] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364782] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364787] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364792] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364797] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364802] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364807] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364811] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364816] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364826] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364830] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364835] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364840] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364844] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364849] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364853] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364858] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364866] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364874] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364880] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364887] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364893] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364900] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364906] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364913] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364920] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364926] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364931] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364936] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364941] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364945] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364950] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364954] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364959] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364963] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364968] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364974] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364979] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364983] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364988] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364992] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.364997] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365002] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365006] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365011] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365015] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365020] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365025] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365029] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365033] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.779 [2024-11-26 19:22:11.365038] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365042] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365047] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365051] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365056] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365060] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365065] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365069] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365074] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365078] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365083] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365087] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365092] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365097] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365103] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365109] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365115] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365120] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365124] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365129] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365134] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365139] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365143] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365148] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365153] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365157] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365161] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 [2024-11-26 19:22:11.365166] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1374370 is same with the state(6) to be set 00:25:58.780 19:22:11 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@45 -- # sleep 3 00:26:02.081 19:22:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:26:02.081 00:26:02.343 19:22:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:26:02.343 [2024-11-26 19:22:14.877351] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877387] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877393] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877398] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877402] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877408] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877412] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877417] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877430] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877435] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877439] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877444] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877448] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877453] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877457] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877462] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877466] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877471] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877475] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877480] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877484] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877489] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877493] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877497] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877502] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877506] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877511] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877515] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877520] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877524] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877529] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877533] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877537] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877542] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877546] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877552] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877556] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877561] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877565] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877570] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877574] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877579] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877583] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877588] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877593] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877597] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877602] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877606] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877611] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877616] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877620] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877625] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877630] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877634] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877638] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877643] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877648] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877652] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877656] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877661] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877665] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.343 [2024-11-26 19:22:14.877670] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877677] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877681] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877686] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877690] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877695] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877701] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877705] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877709] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877714] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877718] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 [2024-11-26 19:22:14.877723] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13751e0 is same with the state(6) to be set 00:26:02.344 19:22:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@50 -- # sleep 3 00:26:05.644 19:22:17 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:05.644 [2024-11-26 19:22:18.066328] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:05.644 19:22:18 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@55 -- # sleep 1 00:26:06.589 19:22:19 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:26:06.848 [2024-11-26 19:22:19.253580] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253616] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253622] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253627] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253632] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253636] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253641] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253645] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253650] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253654] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 [2024-11-26 19:22:19.253659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14c0360 is same with the state(6) to be set 00:26:06.848 19:22:19 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@59 -- # wait 3981653 00:26:13.434 { 00:26:13.434 "results": [ 00:26:13.434 { 00:26:13.434 "job": "NVMe0n1", 00:26:13.434 "core_mask": "0x1", 00:26:13.434 "workload": "verify", 00:26:13.434 "status": "finished", 00:26:13.434 "verify_range": { 00:26:13.434 "start": 0, 00:26:13.434 "length": 16384 00:26:13.434 }, 00:26:13.434 "queue_depth": 128, 00:26:13.434 "io_size": 4096, 00:26:13.434 "runtime": 15.009024, 00:26:13.434 "iops": 11272.951525695475, 00:26:13.434 "mibps": 44.03496689724795, 00:26:13.434 "io_failed": 5213, 00:26:13.434 "io_timeout": 0, 00:26:13.434 "avg_latency_us": 10987.158438230443, 00:26:13.434 "min_latency_us": 508.58666666666664, 00:26:13.434 "max_latency_us": 14964.053333333333 00:26:13.434 } 00:26:13.434 ], 00:26:13.434 "core_count": 1 00:26:13.434 } 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@61 -- # killprocess 3981405 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # '[' -z 3981405 ']' 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@958 -- # kill -0 3981405 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # uname 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3981405 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:13.434 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3981405' 00:26:13.435 killing process with pid 3981405 00:26:13.435 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@973 -- # kill 3981405 00:26:13.435 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@978 -- # wait 3981405 00:26:13.435 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@63 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:26:13.435 [2024-11-26 19:22:08.700488] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:26:13.435 [2024-11-26 19:22:08.700547] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3981405 ] 00:26:13.435 [2024-11-26 19:22:08.777830] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:13.435 [2024-11-26 19:22:08.813685] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:13.435 Running I/O for 15 seconds... 00:26:13.435 11800.00 IOPS, 46.09 MiB/s [2024-11-26T18:22:26.060Z] [2024-11-26 19:22:11.365798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:101048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:101056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:101064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:101072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:101080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:101088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:101096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:101104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:101112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.365984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.365994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:101120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:101128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:101136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:101144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:101152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:101160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:101168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:101176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:101184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:101192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:101200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:101208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:101216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:101224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:101232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:101240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:101248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:101256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:101264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:101272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:101280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:101288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:101296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:101304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:101312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:101320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.435 [2024-11-26 19:22:11.366430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:101328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.435 [2024-11-26 19:22:11.366437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:101336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:101352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:101360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:101368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:101376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:101384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:101392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:101400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:101408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:101416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:101424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:101432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:101440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:101448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:101456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:101464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:101472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:101480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:101488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:101496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:101504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:101512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:101520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:101528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:101536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:101544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:101552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:101560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:101568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:101576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:101584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:101592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.366988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.366997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:101600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.367004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.367013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:101608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.367020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.367030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:101616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.367037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.367046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:101624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.367053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.367062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:101632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.436 [2024-11-26 19:22:11.367070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.436 [2024-11-26 19:22:11.367080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:101640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:101648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:101656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:101664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:101672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:101680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:101688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:101696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.437 [2024-11-26 19:22:11.367202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:101704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:101712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:101720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:101728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:101736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:101744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:101752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:101760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:101768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:101776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:101784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:101792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:101800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:101808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367445] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:101816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:101824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:101832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:101840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:101848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:101856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:101864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:101872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:101880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:101888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:101896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:101904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:101912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:101920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:101928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:101936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:101944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:101952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.437 [2024-11-26 19:22:11.367734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.437 [2024-11-26 19:22:11.367743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:101960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:101968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:101976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:101984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:101992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:102000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:102008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:102016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:102024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:102032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:102040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:102048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:102056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.438 [2024-11-26 19:22:11.367950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.367978] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:26:13.438 [2024-11-26 19:22:11.367985] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:26:13.438 [2024-11-26 19:22:11.367992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102064 len:8 PRP1 0x0 PRP2 0x0 00:26:13.438 [2024-11-26 19:22:11.368001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.368042] bdev_nvme.c:2052:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:26:13.438 [2024-11-26 19:22:11.368063] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.438 [2024-11-26 19:22:11.368071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.368079] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.438 [2024-11-26 19:22:11.368087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.368094] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.438 [2024-11-26 19:22:11.368101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.368109] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.438 [2024-11-26 19:22:11.368116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:11.368133] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:26:13.438 [2024-11-26 19:22:11.371720] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:26:13.438 [2024-11-26 19:22:11.371746] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12b5d90 (9): Bad file descriptor 00:26:13.438 [2024-11-26 19:22:11.437265] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] Resetting controller successful. 00:26:13.438 11297.50 IOPS, 44.13 MiB/s [2024-11-26T18:22:26.063Z] 11521.67 IOPS, 45.01 MiB/s [2024-11-26T18:22:26.063Z] 11502.25 IOPS, 44.93 MiB/s [2024-11-26T18:22:26.063Z] [2024-11-26 19:22:14.878754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:50016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:50024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:50032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:50040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:50048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:50056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:50064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:50072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:50080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:50088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:50096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.878984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:50104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.878992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.879001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:50112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.879008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.879017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:50120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.879024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.879036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:50128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.879043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.879052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:50136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.879059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.879069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:50144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.879076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.879085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:50152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.879092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.438 [2024-11-26 19:22:14.879102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:50160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.438 [2024-11-26 19:22:14.879109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:50168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:50176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:50184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:50192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:50200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:50208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:50216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:50224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:50232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:50240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:50248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:50256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:50264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:50272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:50280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:50288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:50296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:50304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:50312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:50320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:50328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:50336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:50344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:50352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:50360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:50368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:50376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:50384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:50392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:50400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:50408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:50416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:50424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.439 [2024-11-26 19:22:14.879656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.439 [2024-11-26 19:22:14.879667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:50432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:50440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:50448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:50456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:50464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:50472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:50656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.879772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:50480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:50488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:50496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:50504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:50512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:50520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:50528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:50536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:50544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:50552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:50560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:50568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.879989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:50576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.879996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:50584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.880013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:50592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.880029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:50600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.440 [2024-11-26 19:22:14.880046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:50664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:50672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:50680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:50688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:50696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:50704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:50712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:50720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:50728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:50736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:50744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:50752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:50760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:50768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:50776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:50784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:50792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.440 [2024-11-26 19:22:14.880327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.440 [2024-11-26 19:22:14.880336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:50800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:50808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:50816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:50824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:50832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:50840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:50848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:50856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:50864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:50872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:50880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:50888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:50896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:50904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:50912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:50920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:50928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:50936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:50944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:50952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:50960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:50968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:50976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:50984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:50992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:51000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:51008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:51016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:51024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:51032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.441 [2024-11-26 19:22:14.880824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:50608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.441 [2024-11-26 19:22:14.880841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:50616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.441 [2024-11-26 19:22:14.880857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880885] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:26:13.441 [2024-11-26 19:22:14.880894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:50624 len:8 PRP1 0x0 PRP2 0x0 00:26:13.441 [2024-11-26 19:22:14.880901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880912] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:26:13.441 [2024-11-26 19:22:14.880917] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:26:13.441 [2024-11-26 19:22:14.880924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:50632 len:8 PRP1 0x0 PRP2 0x0 00:26:13.441 [2024-11-26 19:22:14.880931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880939] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:26:13.441 [2024-11-26 19:22:14.880945] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:26:13.441 [2024-11-26 19:22:14.880956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:50640 len:8 PRP1 0x0 PRP2 0x0 00:26:13.441 [2024-11-26 19:22:14.880963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.880971] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:26:13.441 [2024-11-26 19:22:14.880976] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:26:13.441 [2024-11-26 19:22:14.880982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:50648 len:8 PRP1 0x0 PRP2 0x0 00:26:13.441 [2024-11-26 19:22:14.880990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.881029] bdev_nvme.c:2052:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] Start failover from 10.0.0.2:4421 to 10.0.0.2:4422 00:26:13.441 [2024-11-26 19:22:14.881050] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.441 [2024-11-26 19:22:14.881059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.441 [2024-11-26 19:22:14.881067] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.442 [2024-11-26 19:22:14.881074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:14.881082] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.442 [2024-11-26 19:22:14.881090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:14.881098] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.442 [2024-11-26 19:22:14.881105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:14.881112] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 3] in failed state. 00:26:13.442 [2024-11-26 19:22:14.884701] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] resetting controller 00:26:13.442 [2024-11-26 19:22:14.884729] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12b5d90 (9): Bad file descriptor 00:26:13.442 [2024-11-26 19:22:14.916125] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] Resetting controller successful. 00:26:13.442 11411.80 IOPS, 44.58 MiB/s [2024-11-26T18:22:26.067Z] 11463.33 IOPS, 44.78 MiB/s [2024-11-26T18:22:26.067Z] 11432.86 IOPS, 44.66 MiB/s [2024-11-26T18:22:26.067Z] 11384.00 IOPS, 44.47 MiB/s [2024-11-26T18:22:26.067Z] [2024-11-26 19:22:19.253827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:58432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.253866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.253883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:58440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.253891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.253901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:58448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.253909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.253919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:58456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.253932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.253942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:59256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.253949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.253959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:59264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.253966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.253976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:59272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.253983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.253992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:59280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.253999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:59288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:59296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:59304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:59312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:59320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:59328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:59336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:59344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:59352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:59360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:59368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:59376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:59384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:59392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.442 [2024-11-26 19:22:19.254229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:58464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:58472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:58480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:58488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:58496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:58504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:58512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:58520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:58528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:58536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:58544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254413] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:58552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.442 [2024-11-26 19:22:19.254439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:58560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.442 [2024-11-26 19:22:19.254446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:58568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:58576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:58584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:58592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:58600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:58608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:58616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:59400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.443 [2024-11-26 19:22:19.254580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:59408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.443 [2024-11-26 19:22:19.254596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:59416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.443 [2024-11-26 19:22:19.254613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:59424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.443 [2024-11-26 19:22:19.254629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:59432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.443 [2024-11-26 19:22:19.254645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:59440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.443 [2024-11-26 19:22:19.254662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:58624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:58632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:58640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:58648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:58656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:58664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:58672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:58680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:58688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:58696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:58704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:58712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:58720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:58728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:58736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:58744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:58752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:58760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:58768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.254986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.254996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:58776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.255003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.255013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:58784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.443 [2024-11-26 19:22:19.255020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.443 [2024-11-26 19:22:19.255030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:58792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:58800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:59448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.444 [2024-11-26 19:22:19.255072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:58808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:58816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:58824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:58832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:58840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:58848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:58856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:58864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:58872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:58880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:58888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:58896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:58904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:58912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:58920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:58928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:58936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:58944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:58952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:58960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:58968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:58976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:58984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:58992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:59000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:59008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:59016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:59024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:59032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:59040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:59048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:59056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:59064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:59072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:59080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:59088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:59096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.444 [2024-11-26 19:22:19.255689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.444 [2024-11-26 19:22:19.255698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:59104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255705] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:59112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:59120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:59128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:59136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:59144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:59152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:59160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:59168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:59176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:59184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:59192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:59200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:59208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:59216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:59224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:59232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.255983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:59240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:13.445 [2024-11-26 19:22:19.255991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.256012] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:26:13.445 [2024-11-26 19:22:19.256018] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:26:13.445 [2024-11-26 19:22:19.256026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:59248 len:8 PRP1 0x0 PRP2 0x0 00:26:13.445 [2024-11-26 19:22:19.256034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.256076] bdev_nvme.c:2052:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] Start failover from 10.0.0.2:4422 to 10.0.0.2:4420 00:26:13.445 [2024-11-26 19:22:19.256097] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.445 [2024-11-26 19:22:19.256108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.256116] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.445 [2024-11-26 19:22:19.256124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.256132] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.445 [2024-11-26 19:22:19.256139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.256147] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:13.445 [2024-11-26 19:22:19.256154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:13.445 [2024-11-26 19:22:19.256162] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 5] in failed state. 00:26:13.445 [2024-11-26 19:22:19.259768] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] resetting controller 00:26:13.445 [2024-11-26 19:22:19.259796] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12b5d90 (9): Bad file descriptor 00:26:13.445 [2024-11-26 19:22:19.285374] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 6] Resetting controller successful. 00:26:13.445 11316.33 IOPS, 44.20 MiB/s [2024-11-26T18:22:26.070Z] 11303.70 IOPS, 44.16 MiB/s [2024-11-26T18:22:26.070Z] 11288.73 IOPS, 44.10 MiB/s [2024-11-26T18:22:26.070Z] 11281.25 IOPS, 44.07 MiB/s [2024-11-26T18:22:26.070Z] 11276.31 IOPS, 44.05 MiB/s [2024-11-26T18:22:26.070Z] 11286.86 IOPS, 44.09 MiB/s 00:26:13.445 Latency(us) 00:26:13.445 [2024-11-26T18:22:26.070Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:13.445 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:26:13.445 Verification LBA range: start 0x0 length 0x4000 00:26:13.445 NVMe0n1 : 15.01 11272.95 44.03 347.32 0.00 10987.16 508.59 14964.05 00:26:13.445 [2024-11-26T18:22:26.070Z] =================================================================================================================== 00:26:13.445 [2024-11-26T18:22:26.070Z] Total : 11272.95 44.03 347.32 0.00 10987.16 508.59 14964.05 00:26:13.445 Received shutdown signal, test time was about 15.000000 seconds 00:26:13.445 00:26:13.445 Latency(us) 00:26:13.445 [2024-11-26T18:22:26.070Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:13.445 [2024-11-26T18:22:26.070Z] =================================================================================================================== 00:26:13.445 [2024-11-26T18:22:26.070Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # grep -c 'Resetting controller successful' 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # count=3 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@67 -- # (( count != 3 )) 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@73 -- # bdevperf_pid=3984516 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@75 -- # waitforlisten 3984516 /var/tmp/bdevperf.sock 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 1 -f 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # '[' -z 3984516 ']' 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:13.445 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:13.445 19:22:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:26:14.017 19:22:26 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:14.018 19:22:26 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@868 -- # return 0 00:26:14.018 19:22:26 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:26:14.018 [2024-11-26 19:22:26.549920] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:26:14.018 19:22:26 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:26:14.287 [2024-11-26 19:22:26.726310] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:26:14.287 19:22:26 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:26:14.616 NVMe0n1 00:26:14.616 19:22:26 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:26:14.927 00:26:14.927 19:22:27 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:26:15.219 00:26:15.219 19:22:27 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:26:15.219 19:22:27 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # grep -q NVMe0 00:26:15.219 19:22:27 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:26:15.480 19:22:27 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@87 -- # sleep 3 00:26:18.780 19:22:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:26:18.780 19:22:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # grep -q NVMe0 00:26:18.780 19:22:31 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@90 -- # run_test_pid=3985719 00:26:18.780 19:22:31 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:26:18.780 19:22:31 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@92 -- # wait 3985719 00:26:19.721 { 00:26:19.721 "results": [ 00:26:19.721 { 00:26:19.721 "job": "NVMe0n1", 00:26:19.721 "core_mask": "0x1", 00:26:19.721 "workload": "verify", 00:26:19.721 "status": "finished", 00:26:19.721 "verify_range": { 00:26:19.721 "start": 0, 00:26:19.721 "length": 16384 00:26:19.721 }, 00:26:19.721 "queue_depth": 128, 00:26:19.721 "io_size": 4096, 00:26:19.721 "runtime": 1.005236, 00:26:19.721 "iops": 11014.328973494781, 00:26:19.721 "mibps": 43.02472255271399, 00:26:19.721 "io_failed": 0, 00:26:19.721 "io_timeout": 0, 00:26:19.721 "avg_latency_us": 11566.919460500963, 00:26:19.721 "min_latency_us": 1051.3066666666666, 00:26:19.721 "max_latency_us": 9994.24 00:26:19.721 } 00:26:19.721 ], 00:26:19.721 "core_count": 1 00:26:19.721 } 00:26:19.721 19:22:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@94 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:26:19.721 [2024-11-26 19:22:25.602435] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:26:19.721 [2024-11-26 19:22:25.602494] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3984516 ] 00:26:19.721 [2024-11-26 19:22:25.680461] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:19.721 [2024-11-26 19:22:25.715959] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:19.721 [2024-11-26 19:22:27.953566] bdev_nvme.c:2052:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 7] Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:26:19.721 [2024-11-26 19:22:27.953611] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:19.721 [2024-11-26 19:22:27.953624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:19.721 [2024-11-26 19:22:27.953633] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:19.721 [2024-11-26 19:22:27.953641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:19.721 [2024-11-26 19:22:27.953649] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:19.721 [2024-11-26 19:22:27.953656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:19.721 [2024-11-26 19:22:27.953664] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:19.721 [2024-11-26 19:22:27.953671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:19.721 [2024-11-26 19:22:27.953679] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 7] in failed state. 00:26:19.721 [2024-11-26 19:22:27.953707] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 7] resetting controller 00:26:19.721 [2024-11-26 19:22:27.953723] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1159d90 (9): Bad file descriptor 00:26:19.721 [2024-11-26 19:22:28.046084] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 10] Resetting controller successful. 00:26:19.721 Running I/O for 1 seconds... 00:26:19.721 10944.00 IOPS, 42.75 MiB/s 00:26:19.721 Latency(us) 00:26:19.721 [2024-11-26T18:22:32.346Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:19.721 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:26:19.721 Verification LBA range: start 0x0 length 0x4000 00:26:19.721 NVMe0n1 : 1.01 11014.33 43.02 0.00 0.00 11566.92 1051.31 9994.24 00:26:19.721 [2024-11-26T18:22:32.347Z] =================================================================================================================== 00:26:19.722 [2024-11-26T18:22:32.347Z] Total : 11014.33 43.02 0.00 0.00 11566.92 1051.31 9994.24 00:26:19.722 19:22:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:26:19.722 19:22:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # grep -q NVMe0 00:26:19.982 19:22:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@98 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:26:20.243 19:22:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:26:20.243 19:22:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # grep -q NVMe0 00:26:20.243 19:22:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:26:20.502 19:22:33 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@101 -- # sleep 3 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # grep -q NVMe0 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@108 -- # killprocess 3984516 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # '[' -z 3984516 ']' 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@958 -- # kill -0 3984516 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # uname 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3984516 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3984516' 00:26:23.804 killing process with pid 3984516 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@973 -- # kill 3984516 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@978 -- # wait 3984516 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@110 -- # sync 00:26:23.804 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:26:24.067 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@113 -- # trap - SIGINT SIGTERM EXIT 00:26:24.067 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@115 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@116 -- # nvmftestfini 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@516 -- # nvmfcleanup 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@121 -- # sync 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@124 -- # set +e 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:24.068 rmmod nvme_tcp 00:26:24.068 rmmod nvme_fabrics 00:26:24.068 rmmod nvme_keyring 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@128 -- # set -e 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@129 -- # return 0 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@517 -- # '[' -n 3980825 ']' 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@518 -- # killprocess 3980825 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # '[' -z 3980825 ']' 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@958 -- # kill -0 3980825 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # uname 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:24.068 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3980825 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3980825' 00:26:24.329 killing process with pid 3980825 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@973 -- # kill 3980825 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@978 -- # wait 3980825 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@297 -- # iptr 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@791 -- # iptables-save 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@791 -- # iptables-restore 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:24.329 19:22:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:26.880 19:22:38 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:26.880 00:26:26.880 real 0m41.311s 00:26:26.880 user 2m3.852s 00:26:26.880 sys 0m9.405s 00:26:26.880 19:22:38 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:26.880 19:22:38 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:26:26.880 ************************************ 00:26:26.880 END TEST nvmf_failover 00:26:26.880 ************************************ 00:26:26.880 19:22:38 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@26 -- # run_test nvmf_host_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:26:26.880 19:22:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:26.880 19:22:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:26.880 19:22:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:26:26.880 ************************************ 00:26:26.880 START TEST nvmf_host_discovery 00:26:26.880 ************************************ 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:26:26.880 * Looking for test storage... 00:26:26.880 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1693 -- # lcov --version 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:26.880 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@344 -- # case "$op" in 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@345 -- # : 1 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # decimal 1 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=1 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 1 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # decimal 2 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=2 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 2 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # return 0 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:26:26.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:26.881 --rc genhtml_branch_coverage=1 00:26:26.881 --rc genhtml_function_coverage=1 00:26:26.881 --rc genhtml_legend=1 00:26:26.881 --rc geninfo_all_blocks=1 00:26:26.881 --rc geninfo_unexecuted_blocks=1 00:26:26.881 00:26:26.881 ' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:26:26.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:26.881 --rc genhtml_branch_coverage=1 00:26:26.881 --rc genhtml_function_coverage=1 00:26:26.881 --rc genhtml_legend=1 00:26:26.881 --rc geninfo_all_blocks=1 00:26:26.881 --rc geninfo_unexecuted_blocks=1 00:26:26.881 00:26:26.881 ' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:26:26.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:26.881 --rc genhtml_branch_coverage=1 00:26:26.881 --rc genhtml_function_coverage=1 00:26:26.881 --rc genhtml_legend=1 00:26:26.881 --rc geninfo_all_blocks=1 00:26:26.881 --rc geninfo_unexecuted_blocks=1 00:26:26.881 00:26:26.881 ' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:26:26.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:26.881 --rc genhtml_branch_coverage=1 00:26:26.881 --rc genhtml_function_coverage=1 00:26:26.881 --rc genhtml_legend=1 00:26:26.881 --rc geninfo_all_blocks=1 00:26:26.881 --rc geninfo_unexecuted_blocks=1 00:26:26.881 00:26:26.881 ' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # uname -s 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@5 -- # export PATH 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@51 -- # : 0 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:26:26.881 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@11 -- # '[' tcp == rdma ']' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@16 -- # DISCOVERY_PORT=8009 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@17 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@20 -- # NQN=nqn.2016-06.io.spdk:cnode 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@22 -- # HOST_NQN=nqn.2021-12.io.spdk:test 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@23 -- # HOST_SOCK=/tmp/host.sock 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@25 -- # nvmftestinit 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@476 -- # prepare_net_devs 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@438 -- # local -g is_hw=no 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # remove_spdk_ns 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:26:26.881 19:22:39 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # e810=() 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # x722=() 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # mlx=() 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:26:35.029 Found 0000:31:00.0 (0x8086 - 0x159b) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:26:35.029 Found 0000:31:00.1 (0x8086 - 0x159b) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:26:35.029 Found net devices under 0000:31:00.0: cvl_0_0 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:26:35.029 Found net devices under 0000:31:00.1: cvl_0_1 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # is_hw=yes 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:35.029 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:35.030 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:35.030 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.656 ms 00:26:35.030 00:26:35.030 --- 10.0.0.2 ping statistics --- 00:26:35.030 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:35.030 rtt min/avg/max/mdev = 0.656/0.656/0.656/0.000 ms 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:35.030 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:35.030 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.325 ms 00:26:35.030 00:26:35.030 --- 10.0.0.1 ping statistics --- 00:26:35.030 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:35.030 rtt min/avg/max/mdev = 0.325/0.325/0.325/0.000 ms 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@450 -- # return 0 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@30 -- # nvmfappstart -m 0x2 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@726 -- # xtrace_disable 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@509 -- # nvmfpid=3991516 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@510 -- # waitforlisten 3991516 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # '[' -z 3991516 ']' 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:35.030 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:35.030 19:22:47 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:35.291 [2024-11-26 19:22:47.656822] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:26:35.291 [2024-11-26 19:22:47.656900] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:35.291 [2024-11-26 19:22:47.762044] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:35.291 [2024-11-26 19:22:47.812322] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:35.291 [2024-11-26 19:22:47.812375] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:35.291 [2024-11-26 19:22:47.812383] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:35.291 [2024-11-26 19:22:47.812391] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:35.291 [2024-11-26 19:22:47.812397] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:35.291 [2024-11-26 19:22:47.813229] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:35.864 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:35.864 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@868 -- # return 0 00:26:35.864 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:26:35.864 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@732 -- # xtrace_disable 00:26:35.864 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@32 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.126 [2024-11-26 19:22:48.527551] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2014-08.org.nvmexpress.discovery -t tcp -a 10.0.0.2 -s 8009 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.126 [2024-11-26 19:22:48.539815] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@35 -- # rpc_cmd bdev_null_create null0 1000 512 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.126 null0 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@36 -- # rpc_cmd bdev_null_create null1 1000 512 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.126 null1 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@37 -- # rpc_cmd bdev_wait_for_examine 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@45 -- # hostpid=3991601 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@46 -- # waitforlisten 3991601 /tmp/host.sock 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # '[' -z 3991601 ']' 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@839 -- # local rpc_addr=/tmp/host.sock 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:26:36.126 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:36.126 19:22:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.126 [2024-11-26 19:22:48.640089] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:26:36.126 [2024-11-26 19:22:48.640155] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3991601 ] 00:26:36.126 [2024-11-26 19:22:48.723270] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:36.387 [2024-11-26 19:22:48.765172] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@868 -- # return 0 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@48 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@50 -- # rpc_cmd -s /tmp/host.sock log_set_flag bdev_nvme 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@51 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@72 -- # notify_id=0 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # get_subsystem_names 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # [[ '' == '' ]] 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # get_bdev_list 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # [[ '' == '' ]] 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@86 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # get_subsystem_names 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:36.957 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # [[ '' == '' ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # get_bdev_list 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # [[ '' == '' ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@90 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # get_subsystem_names 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # [[ '' == '' ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # get_bdev_list 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # [[ '' == '' ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@96 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.219 [2024-11-26 19:22:49.786965] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # get_subsystem_names 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:37.219 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # [[ '' == '' ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # get_bdev_list 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # [[ '' == '' ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@99 -- # is_notification_count_eq 0 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=0 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@103 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2021-12.io.spdk:test 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@105 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ '' == \n\v\m\e\0 ]] 00:26:37.480 19:22:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@924 -- # sleep 1 00:26:38.051 [2024-11-26 19:22:50.507984] bdev_nvme.c:7484:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:26:38.051 [2024-11-26 19:22:50.508006] bdev_nvme.c:7570:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:26:38.051 [2024-11-26 19:22:50.508020] bdev_nvme.c:7447:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:26:38.051 [2024-11-26 19:22:50.596315] bdev_nvme.c:7413:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:26:38.312 [2024-11-26 19:22:50.818607] bdev_nvme.c:5636:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr was created to 10.0.0.2:4420 00:26:38.312 [2024-11-26 19:22:50.819624] bdev_nvme.c:1985:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x16136a0:1 started. 00:26:38.312 [2024-11-26 19:22:50.821244] bdev_nvme.c:7303:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:26:38.313 [2024-11-26 19:22:50.821262] bdev_nvme.c:7262:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:26:38.313 [2024-11-26 19:22:50.825415] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x16136a0 was disconnected and freed. delete nvme_qpair. 00:26:38.573 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.574 19:22:50 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@106 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1"' ']]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 == \n\v\m\e\0\n\1 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@107 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT"' ']]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4420 == \4\4\2\0 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@108 -- # is_notification_count_eq 1 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=1 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@111 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null1 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@113 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:26:38.574 [2024-11-26 19:22:51.192888] bdev_nvme.c:1985:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x1613a20:1 started. 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.574 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:38.574 [2024-11-26 19:22:51.195652] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x1613a20 was disconnected and freed. delete nvme_qpair. 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@114 -- # is_notification_count_eq 1 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:26:38.835 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 1 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@118 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.836 [2024-11-26 19:22:51.294867] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:26:38.836 [2024-11-26 19:22:51.295043] bdev_nvme.c:7466:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:26:38.836 [2024-11-26 19:22:51.295063] bdev_nvme.c:7447:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@120 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@121 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:38.836 [2024-11-26 19:22:51.381319] bdev_nvme.c:7408:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new path for nvme0 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@122 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4420 == \4\4\2\0\ \4\4\2\1 ]] 00:26:38.836 19:22:51 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@924 -- # sleep 1 00:26:39.097 [2024-11-26 19:22:51.480193] bdev_nvme.c:5636:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] ctrlr was created to 10.0.0.2:4421 00:26:39.097 [2024-11-26 19:22:51.480228] bdev_nvme.c:7303:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:26:39.097 [2024-11-26 19:22:51.480235] bdev_nvme.c:7262:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:26:39.097 [2024-11-26 19:22:51.480245] bdev_nvme.c:7262:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4420 4421 == \4\4\2\0\ \4\4\2\1 ]] 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@123 -- # is_notification_count_eq 0 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.040 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@127 -- # rpc_cmd nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.041 [2024-11-26 19:22:52.542253] bdev_nvme.c:7466:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:26:40.041 [2024-11-26 19:22:52.542276] bdev_nvme.c:7447:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:26:40.041 [2024-11-26 19:22:52.542855] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:40.041 [2024-11-26 19:22:52.542877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:40.041 [2024-11-26 19:22:52.542887] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:40.041 [2024-11-26 19:22:52.542894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:40.041 [2024-11-26 19:22:52.542902] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:40.041 [2024-11-26 19:22:52.542910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:40.041 [2024-11-26 19:22:52.542918] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:40.041 [2024-11-26 19:22:52.542925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:40.041 [2024-11-26 19:22:52.542933] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@129 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.041 [2024-11-26 19:22:52.552870] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.041 [2024-11-26 19:22:52.562900] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:26:40.041 [2024-11-26 19:22:52.562914] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:26:40.041 [2024-11-26 19:22:52.562919] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.562924] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:26:40.041 [2024-11-26 19:22:52.562942] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.563326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:40.041 [2024-11-26 19:22:52.563340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x15e3d90 with addr=10.0.0.2, port=4420 00:26:40.041 [2024-11-26 19:22:52.563349] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.041 [2024-11-26 19:22:52.563364] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.041 [2024-11-26 19:22:52.563382] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:26:40.041 [2024-11-26 19:22:52.563390] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:26:40.041 [2024-11-26 19:22:52.563399] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:26:40.041 [2024-11-26 19:22:52.563406] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:26:40.041 [2024-11-26 19:22:52.563412] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:26:40.041 [2024-11-26 19:22:52.563416] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:26:40.041 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.041 [2024-11-26 19:22:52.572973] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:26:40.041 [2024-11-26 19:22:52.572985] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:26:40.041 [2024-11-26 19:22:52.572989] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.572994] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:26:40.041 [2024-11-26 19:22:52.573008] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.573232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:40.041 [2024-11-26 19:22:52.573244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x15e3d90 with addr=10.0.0.2, port=4420 00:26:40.041 [2024-11-26 19:22:52.573252] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.041 [2024-11-26 19:22:52.573263] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.041 [2024-11-26 19:22:52.573274] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:26:40.041 [2024-11-26 19:22:52.573281] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:26:40.041 [2024-11-26 19:22:52.573288] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:26:40.041 [2024-11-26 19:22:52.573294] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:26:40.041 [2024-11-26 19:22:52.573299] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:26:40.041 [2024-11-26 19:22:52.573304] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:26:40.041 [2024-11-26 19:22:52.583040] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:26:40.041 [2024-11-26 19:22:52.583052] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:26:40.041 [2024-11-26 19:22:52.583057] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.583061] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:26:40.041 [2024-11-26 19:22:52.583075] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.583364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:40.041 [2024-11-26 19:22:52.583378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x15e3d90 with addr=10.0.0.2, port=4420 00:26:40.041 [2024-11-26 19:22:52.583385] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.041 [2024-11-26 19:22:52.583396] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.041 [2024-11-26 19:22:52.583407] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:26:40.041 [2024-11-26 19:22:52.583413] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:26:40.041 [2024-11-26 19:22:52.583421] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:26:40.041 [2024-11-26 19:22:52.583427] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:26:40.041 [2024-11-26 19:22:52.583431] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:26:40.041 [2024-11-26 19:22:52.583436] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:26:40.041 [2024-11-26 19:22:52.593106] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:26:40.041 [2024-11-26 19:22:52.593120] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:26:40.041 [2024-11-26 19:22:52.593125] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.593129] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:26:40.041 [2024-11-26 19:22:52.593145] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:26:40.041 [2024-11-26 19:22:52.593433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:40.041 [2024-11-26 19:22:52.593446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x15e3d90 with addr=10.0.0.2, port=4420 00:26:40.042 [2024-11-26 19:22:52.593456] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.042 [2024-11-26 19:22:52.593467] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.042 [2024-11-26 19:22:52.593477] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:26:40.042 [2024-11-26 19:22:52.593484] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:26:40.042 [2024-11-26 19:22:52.593492] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:26:40.042 [2024-11-26 19:22:52.593499] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:26:40.042 [2024-11-26 19:22:52.593504] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:26:40.042 [2024-11-26 19:22:52.593508] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@130 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:40.042 [2024-11-26 19:22:52.603176] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:26:40.042 [2024-11-26 19:22:52.603188] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:26:40.042 [2024-11-26 19:22:52.603193] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:26:40.042 [2024-11-26 19:22:52.603198] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:26:40.042 [2024-11-26 19:22:52.603212] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:40.042 [2024-11-26 19:22:52.603415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:40.042 [2024-11-26 19:22:52.603429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x15e3d90 with addr=10.0.0.2, port=4420 00:26:40.042 [2024-11-26 19:22:52.603437] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.042 [2024-11-26 19:22:52.603449] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.042 [2024-11-26 19:22:52.603460] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:26:40.042 [2024-11-26 19:22:52.603467] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:26:40.042 [2024-11-26 19:22:52.603475] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:26:40.042 [2024-11-26 19:22:52.603481] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:26:40.042 [2024-11-26 19:22:52.603486] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:26:40.042 [2024-11-26 19:22:52.603493] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:40.042 [2024-11-26 19:22:52.613244] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:26:40.042 [2024-11-26 19:22:52.613258] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:26:40.042 [2024-11-26 19:22:52.613263] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:26:40.042 [2024-11-26 19:22:52.613267] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:26:40.042 [2024-11-26 19:22:52.613282] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:26:40.042 [2024-11-26 19:22:52.613570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:40.042 [2024-11-26 19:22:52.613582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x15e3d90 with addr=10.0.0.2, port=4420 00:26:40.042 [2024-11-26 19:22:52.613589] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.042 [2024-11-26 19:22:52.613605] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.042 [2024-11-26 19:22:52.613616] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:26:40.042 [2024-11-26 19:22:52.613622] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:26:40.042 [2024-11-26 19:22:52.613630] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:26:40.042 [2024-11-26 19:22:52.613636] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:26:40.042 [2024-11-26 19:22:52.613641] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:26:40.042 [2024-11-26 19:22:52.613646] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:26:40.042 [2024-11-26 19:22:52.623313] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:26:40.042 [2024-11-26 19:22:52.623325] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:26:40.042 [2024-11-26 19:22:52.623330] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:26:40.042 [2024-11-26 19:22:52.623335] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:26:40.042 [2024-11-26 19:22:52.623348] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:26:40.042 [2024-11-26 19:22:52.623638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:40.042 [2024-11-26 19:22:52.623649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x15e3d90 with addr=10.0.0.2, port=4420 00:26:40.042 [2024-11-26 19:22:52.623656] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e3d90 is same with the state(6) to be set 00:26:40.042 [2024-11-26 19:22:52.623667] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15e3d90 (9): Bad file descriptor 00:26:40.042 [2024-11-26 19:22:52.623678] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:26:40.042 [2024-11-26 19:22:52.623685] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:26:40.042 [2024-11-26 19:22:52.623692] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:26:40.042 [2024-11-26 19:22:52.623698] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:26:40.042 [2024-11-26 19:22:52.623703] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:26:40.042 [2024-11-26 19:22:52.623708] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.042 [2024-11-26 19:22:52.627980] bdev_nvme.c:7271:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 not found 00:26:40.042 [2024-11-26 19:22:52.627998] bdev_nvme.c:7262:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@131 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_SECOND_PORT"' ']]' 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.042 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4421 == \4\4\2\1 ]] 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@132 -- # is_notification_count_eq 0 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@134 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_stop_discovery -b nvme 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.303 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@136 -- # waitforcondition '[[ "$(get_subsystem_names)" == "" ]]' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "" ]]' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '""' ']]' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ '' == '' ]] 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@137 -- # waitforcondition '[[ "$(get_bdev_list)" == "" ]]' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "" ]]' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '""' ']]' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ '' == '' ]] 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@138 -- # is_notification_count_eq 2 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=2 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=2 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=4 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@141 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.304 19:22:52 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:41.687 [2024-11-26 19:22:53.956824] bdev_nvme.c:7484:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:26:41.687 [2024-11-26 19:22:53.956842] bdev_nvme.c:7570:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:26:41.687 [2024-11-26 19:22:53.956854] bdev_nvme.c:7447:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:26:41.687 [2024-11-26 19:22:54.085282] bdev_nvme.c:7413:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new subsystem nvme0 00:26:41.949 [2024-11-26 19:22:54.349647] bdev_nvme.c:5636:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] ctrlr was created to 10.0.0.2:4421 00:26:41.949 [2024-11-26 19:22:54.350399] bdev_nvme.c:1985:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] Connecting qpair 0x160d180:1 started. 00:26:41.949 [2024-11-26 19:22:54.352178] bdev_nvme.c:7303:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:26:41.949 [2024-11-26 19:22:54.352205] bdev_nvme.c:7262:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@143 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:26:41.949 [2024-11-26 19:22:54.354615] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] qpair 0x160d180 was disconnected and freed. delete nvme_qpair. 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # local es=0 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:41.949 request: 00:26:41.949 { 00:26:41.949 "name": "nvme", 00:26:41.949 "trtype": "tcp", 00:26:41.949 "traddr": "10.0.0.2", 00:26:41.949 "adrfam": "ipv4", 00:26:41.949 "trsvcid": "8009", 00:26:41.949 "hostnqn": "nqn.2021-12.io.spdk:test", 00:26:41.949 "wait_for_attach": true, 00:26:41.949 "method": "bdev_nvme_start_discovery", 00:26:41.949 "req_id": 1 00:26:41.949 } 00:26:41.949 Got JSON-RPC error response 00:26:41.949 response: 00:26:41.949 { 00:26:41.949 "code": -17, 00:26:41.949 "message": "File exists" 00:26:41.949 } 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # es=1 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # get_discovery_ctrlrs 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # [[ nvme == \n\v\m\e ]] 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # get_bdev_list 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@149 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # local es=0 00:26:41.949 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:41.950 request: 00:26:41.950 { 00:26:41.950 "name": "nvme_second", 00:26:41.950 "trtype": "tcp", 00:26:41.950 "traddr": "10.0.0.2", 00:26:41.950 "adrfam": "ipv4", 00:26:41.950 "trsvcid": "8009", 00:26:41.950 "hostnqn": "nqn.2021-12.io.spdk:test", 00:26:41.950 "wait_for_attach": true, 00:26:41.950 "method": "bdev_nvme_start_discovery", 00:26:41.950 "req_id": 1 00:26:41.950 } 00:26:41.950 Got JSON-RPC error response 00:26:41.950 response: 00:26:41.950 { 00:26:41.950 "code": -17, 00:26:41.950 "message": "File exists" 00:26:41.950 } 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # es=1 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # get_discovery_ctrlrs 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # [[ nvme == \n\v\m\e ]] 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # get_bdev_list 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:41.950 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@155 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # local es=0 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.210 19:22:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:43.153 [2024-11-26 19:22:55.616200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:43.153 [2024-11-26 19:22:55.616227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16066b0 with addr=10.0.0.2, port=8010 00:26:43.153 [2024-11-26 19:22:55.616245] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:26:43.153 [2024-11-26 19:22:55.616252] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:26:43.153 [2024-11-26 19:22:55.616259] bdev_nvme.c:7552:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:26:44.093 [2024-11-26 19:22:56.618662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:44.093 [2024-11-26 19:22:56.618687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16066b0 with addr=10.0.0.2, port=8010 00:26:44.093 [2024-11-26 19:22:56.618699] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:26:44.093 [2024-11-26 19:22:56.618706] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:26:44.093 [2024-11-26 19:22:56.618712] bdev_nvme.c:7552:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:26:45.038 [2024-11-26 19:22:57.620653] bdev_nvme.c:7527:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] timed out while attaching discovery ctrlr 00:26:45.038 request: 00:26:45.038 { 00:26:45.038 "name": "nvme_second", 00:26:45.038 "trtype": "tcp", 00:26:45.038 "traddr": "10.0.0.2", 00:26:45.038 "adrfam": "ipv4", 00:26:45.038 "trsvcid": "8010", 00:26:45.038 "hostnqn": "nqn.2021-12.io.spdk:test", 00:26:45.038 "wait_for_attach": false, 00:26:45.038 "attach_timeout_ms": 3000, 00:26:45.038 "method": "bdev_nvme_start_discovery", 00:26:45.038 "req_id": 1 00:26:45.038 } 00:26:45.038 Got JSON-RPC error response 00:26:45.038 response: 00:26:45.038 { 00:26:45.038 "code": -110, 00:26:45.038 "message": "Connection timed out" 00:26:45.038 } 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # es=1 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # get_discovery_ctrlrs 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:26:45.038 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # [[ nvme == \n\v\m\e ]] 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@159 -- # trap - SIGINT SIGTERM EXIT 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@161 -- # kill 3991601 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@162 -- # nvmftestfini 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@516 -- # nvmfcleanup 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@121 -- # sync 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@124 -- # set +e 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:45.300 rmmod nvme_tcp 00:26:45.300 rmmod nvme_fabrics 00:26:45.300 rmmod nvme_keyring 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@128 -- # set -e 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@129 -- # return 0 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@517 -- # '[' -n 3991516 ']' 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@518 -- # killprocess 3991516 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@954 -- # '[' -z 3991516 ']' 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@958 -- # kill -0 3991516 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@959 -- # uname 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3991516 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3991516' 00:26:45.300 killing process with pid 3991516 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@973 -- # kill 3991516 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@978 -- # wait 3991516 00:26:45.300 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@297 -- # iptr 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@791 -- # iptables-save 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@791 -- # iptables-restore 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:45.562 19:22:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:47.476 00:26:47.476 real 0m20.996s 00:26:47.476 user 0m23.478s 00:26:47.476 sys 0m7.719s 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:26:47.476 ************************************ 00:26:47.476 END TEST nvmf_host_discovery 00:26:47.476 ************************************ 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@27 -- # run_test nvmf_host_multipath_status /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:26:47.476 ************************************ 00:26:47.476 START TEST nvmf_host_multipath_status 00:26:47.476 ************************************ 00:26:47.476 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:26:47.737 * Looking for test storage... 00:26:47.737 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1693 -- # lcov --version 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # IFS=.-: 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # read -ra ver1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # IFS=.-: 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # read -ra ver2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@338 -- # local 'op=<' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@340 -- # ver1_l=2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@341 -- # ver2_l=1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@344 -- # case "$op" in 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@345 -- # : 1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # decimal 1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # ver1[v]=1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # decimal 2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # ver2[v]=2 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # return 0 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:26:47.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:47.737 --rc genhtml_branch_coverage=1 00:26:47.737 --rc genhtml_function_coverage=1 00:26:47.737 --rc genhtml_legend=1 00:26:47.737 --rc geninfo_all_blocks=1 00:26:47.737 --rc geninfo_unexecuted_blocks=1 00:26:47.737 00:26:47.737 ' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:26:47.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:47.737 --rc genhtml_branch_coverage=1 00:26:47.737 --rc genhtml_function_coverage=1 00:26:47.737 --rc genhtml_legend=1 00:26:47.737 --rc geninfo_all_blocks=1 00:26:47.737 --rc geninfo_unexecuted_blocks=1 00:26:47.737 00:26:47.737 ' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:26:47.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:47.737 --rc genhtml_branch_coverage=1 00:26:47.737 --rc genhtml_function_coverage=1 00:26:47.737 --rc genhtml_legend=1 00:26:47.737 --rc geninfo_all_blocks=1 00:26:47.737 --rc geninfo_unexecuted_blocks=1 00:26:47.737 00:26:47.737 ' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:26:47.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:47.737 --rc genhtml_branch_coverage=1 00:26:47.737 --rc genhtml_function_coverage=1 00:26:47.737 --rc genhtml_legend=1 00:26:47.737 --rc geninfo_all_blocks=1 00:26:47.737 --rc geninfo_unexecuted_blocks=1 00:26:47.737 00:26:47.737 ' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # uname -s 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@15 -- # shopt -s extglob 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:47.737 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@5 -- # export PATH 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@51 -- # : 0 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:26:47.738 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@55 -- # have_pci_nics=0 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@12 -- # MALLOC_BDEV_SIZE=64 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@16 -- # bpf_sh=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/bpftrace.sh 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@18 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@21 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@31 -- # nvmftestinit 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@476 -- # prepare_net_devs 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@438 -- # local -g is_hw=no 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # remove_spdk_ns 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@309 -- # xtrace_disable 00:26:47.738 19:23:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # pci_devs=() 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # net_devs=() 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:57.757 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # e810=() 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # local -ga e810 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # x722=() 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # local -ga x722 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # mlx=() 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # local -ga mlx 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:26:57.758 Found 0000:31:00.0 (0x8086 - 0x159b) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:26:57.758 Found 0000:31:00.1 (0x8086 - 0x159b) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@418 -- # [[ up == up ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:26:57.758 Found net devices under 0000:31:00.0: cvl_0_0 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@418 -- # [[ up == up ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:26:57.758 Found net devices under 0000:31:00.1: cvl_0_1 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # is_hw=yes 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:57.758 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:57.758 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.615 ms 00:26:57.758 00:26:57.758 --- 10.0.0.2 ping statistics --- 00:26:57.758 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:57.758 rtt min/avg/max/mdev = 0.615/0.615/0.615/0.000 ms 00:26:57.758 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:57.758 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:57.758 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.283 ms 00:26:57.758 00:26:57.759 --- 10.0.0.1 ping statistics --- 00:26:57.759 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:57.759 rtt min/avg/max/mdev = 0.283/0.283/0.283/0.000 ms 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@450 -- # return 0 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@33 -- # nvmfappstart -m 0x3 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@726 -- # xtrace_disable 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@509 -- # nvmfpid=3998396 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@510 -- # waitforlisten 3998396 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # '[' -z 3998396 ']' 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:57.759 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:57.759 19:23:08 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:26:57.759 [2024-11-26 19:23:08.977670] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:26:57.759 [2024-11-26 19:23:08.977740] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:57.759 [2024-11-26 19:23:09.068382] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:26:57.759 [2024-11-26 19:23:09.108590] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:57.759 [2024-11-26 19:23:09.108627] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:57.759 [2024-11-26 19:23:09.108635] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:57.759 [2024-11-26 19:23:09.108642] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:57.759 [2024-11-26 19:23:09.108648] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:57.759 [2024-11-26 19:23:09.109948] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:57.759 [2024-11-26 19:23:09.110103] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@868 -- # return 0 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@732 -- # xtrace_disable 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@34 -- # nvmfapp_pid=3998396 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:26:57.759 [2024-11-26 19:23:09.973417] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:57.759 19:23:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:26:57.759 Malloc0 00:26:57.759 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -r -m 2 00:26:57.759 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:26:58.022 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:58.283 [2024-11-26 19:23:10.665260] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:58.283 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:26:58.284 [2024-11-26 19:23:10.833676] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@45 -- # bdevperf_pid=3998757 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 90 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@47 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@48 -- # waitforlisten 3998757 /var/tmp/bdevperf.sock 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # '[' -z 3998757 ']' 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:58.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:58.284 19:23:10 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:26:58.545 19:23:11 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:58.545 19:23:11 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@868 -- # return 0 00:26:58.545 19:23:11 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_options -r -1 00:26:58.806 19:23:11 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:26:59.377 Nvme0n1 00:26:59.377 19:23:11 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:26:59.638 Nvme0n1 00:26:59.638 19:23:12 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@78 -- # sleep 2 00:26:59.638 19:23:12 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 120 -s /var/tmp/bdevperf.sock perform_tests 00:27:02.185 19:23:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@90 -- # set_ANA_state optimized optimized 00:27:02.185 19:23:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:27:02.185 19:23:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:27:02.185 19:23:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@91 -- # sleep 1 00:27:03.127 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@92 -- # check_status true false true true true true 00:27:03.127 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:27:03.128 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:03.128 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:03.388 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:03.388 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:27:03.388 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:03.388 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:03.388 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:03.388 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:03.388 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:03.389 19:23:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:03.649 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:03.649 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:03.649 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:03.649 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:03.910 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:04.170 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:04.170 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@94 -- # set_ANA_state non_optimized optimized 00:27:04.170 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:27:04.430 19:23:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:27:04.690 19:23:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@95 -- # sleep 1 00:27:05.630 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@96 -- # check_status false true true true true true 00:27:05.630 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:27:05.630 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:05.630 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:05.890 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:06.151 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:06.151 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:06.151 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:06.151 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:06.411 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:06.411 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:06.411 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:06.411 19:23:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:06.411 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:06.411 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:27:06.411 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:06.411 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:06.671 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:06.671 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@100 -- # set_ANA_state non_optimized non_optimized 00:27:06.671 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:27:06.931 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:27:07.191 19:23:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@101 -- # sleep 1 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@102 -- # check_status true false true true true true 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:08.135 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:08.395 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:08.395 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:08.395 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:08.395 19:23:20 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:08.655 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:08.916 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:08.916 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:27:08.916 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:08.916 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:09.177 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:09.177 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@104 -- # set_ANA_state non_optimized inaccessible 00:27:09.177 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:27:09.437 19:23:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:27:09.437 19:23:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@105 -- # sleep 1 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@106 -- # check_status true false true true true false 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:10.820 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:11.118 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:11.118 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:11.118 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:11.118 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:11.402 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:11.402 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:11.403 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:11.403 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:11.403 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:11.403 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:27:11.403 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:11.403 19:23:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:11.691 19:23:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:11.691 19:23:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@108 -- # set_ANA_state inaccessible inaccessible 00:27:11.691 19:23:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:27:11.691 19:23:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:27:11.952 19:23:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@109 -- # sleep 1 00:27:12.893 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@110 -- # check_status false false true true false false 00:27:12.893 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:27:12.893 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:12.893 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:13.153 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:13.153 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:27:13.153 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:13.153 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:13.413 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:13.413 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:13.413 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:13.413 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:13.413 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:13.414 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:13.414 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:13.414 19:23:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:13.674 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:13.674 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:27:13.674 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:13.674 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:13.935 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:13.935 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:27:13.935 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:13.935 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:13.935 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:13.935 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@112 -- # set_ANA_state inaccessible optimized 00:27:13.935 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:27:14.196 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:27:14.456 19:23:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@113 -- # sleep 1 00:27:15.398 19:23:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@114 -- # check_status false true true true false true 00:27:15.398 19:23:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:27:15.398 19:23:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:15.398 19:23:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:15.659 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:15.919 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:15.919 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:15.919 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:15.919 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:16.180 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:16.441 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:16.441 19:23:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_multipath_policy -b Nvme0n1 -p active_active 00:27:16.703 19:23:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@119 -- # set_ANA_state optimized optimized 00:27:16.703 19:23:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:27:16.964 19:23:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:27:16.964 19:23:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@120 -- # sleep 1 00:27:17.905 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@121 -- # check_status true true true true true true 00:27:17.905 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:27:17.905 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:17.905 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:18.166 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:18.166 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:27:18.166 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:18.166 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:18.427 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:18.427 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:18.427 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:18.427 19:23:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:18.688 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:18.949 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:18.949 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:27:18.949 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:18.949 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:19.210 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:19.210 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@123 -- # set_ANA_state non_optimized optimized 00:27:19.210 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:27:19.210 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:27:19.471 19:23:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@124 -- # sleep 1 00:27:20.414 19:23:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@125 -- # check_status false true true true true true 00:27:20.414 19:23:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:27:20.414 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:20.414 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:20.675 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:20.675 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:27:20.675 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:20.675 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:20.934 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:20.934 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:20.934 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:20.934 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:21.194 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:21.453 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:21.453 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:27:21.453 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:21.453 19:23:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:21.713 19:23:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:21.713 19:23:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@129 -- # set_ANA_state non_optimized non_optimized 00:27:21.713 19:23:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:27:21.713 19:23:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:27:21.974 19:23:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@130 -- # sleep 1 00:27:22.916 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@131 -- # check_status true true true true true true 00:27:22.916 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:27:22.916 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:22.916 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:23.178 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:23.178 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:27:23.178 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:23.178 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:23.439 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:23.439 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:23.439 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:23.439 19:23:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:23.439 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:23.439 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:23.439 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:23.439 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:23.701 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:23.701 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:23.701 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:23.701 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:23.962 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:23.962 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:27:23.962 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:23.962 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:24.222 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:24.222 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@133 -- # set_ANA_state non_optimized inaccessible 00:27:24.222 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:27:24.222 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:27:24.482 19:23:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@134 -- # sleep 1 00:27:25.424 19:23:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@135 -- # check_status true false true true true false 00:27:25.424 19:23:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:27:25.424 19:23:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:25.424 19:23:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:27:25.685 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:25.685 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:27:25.685 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:25.685 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:27:25.946 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:25.946 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:27:25.946 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:25.946 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:27:25.947 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:25.947 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:27:25.947 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:25.947 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:27:26.207 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:26.207 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:27:26.207 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:27:26.207 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:26.468 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:27:26.468 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:27:26.468 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:27:26.468 19:23:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:27:26.468 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:27:26.468 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@137 -- # killprocess 3998757 00:27:26.468 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # '[' -z 3998757 ']' 00:27:26.468 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@958 -- # kill -0 3998757 00:27:26.468 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # uname 00:27:26.468 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:26.468 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3998757 00:27:26.729 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:27:26.729 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:27:26.729 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3998757' 00:27:26.729 killing process with pid 3998757 00:27:26.729 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@973 -- # kill 3998757 00:27:26.729 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@978 -- # wait 3998757 00:27:26.729 { 00:27:26.729 "results": [ 00:27:26.729 { 00:27:26.729 "job": "Nvme0n1", 00:27:26.729 "core_mask": "0x4", 00:27:26.729 "workload": "verify", 00:27:26.729 "status": "terminated", 00:27:26.729 "verify_range": { 00:27:26.729 "start": 0, 00:27:26.729 "length": 16384 00:27:26.729 }, 00:27:26.729 "queue_depth": 128, 00:27:26.729 "io_size": 4096, 00:27:26.729 "runtime": 26.768625, 00:27:26.729 "iops": 10707.049764416364, 00:27:26.729 "mibps": 41.82441314225142, 00:27:26.729 "io_failed": 0, 00:27:26.729 "io_timeout": 0, 00:27:26.729 "avg_latency_us": 11936.966275779534, 00:27:26.729 "min_latency_us": 314.02666666666664, 00:27:26.729 "max_latency_us": 3019898.88 00:27:26.729 } 00:27:26.729 ], 00:27:26.729 "core_count": 1 00:27:26.729 } 00:27:26.729 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@139 -- # wait 3998757 00:27:26.729 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@141 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:27:26.729 [2024-11-26 19:23:10.907892] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:27:26.729 [2024-11-26 19:23:10.907959] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3998757 ] 00:27:26.729 [2024-11-26 19:23:10.972438] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:26.729 [2024-11-26 19:23:11.001293] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:27:26.729 Running I/O for 90 seconds... 00:27:26.729 9411.00 IOPS, 36.76 MiB/s [2024-11-26T18:23:39.354Z] 9497.50 IOPS, 37.10 MiB/s [2024-11-26T18:23:39.354Z] 9493.67 IOPS, 37.08 MiB/s [2024-11-26T18:23:39.354Z] 9522.00 IOPS, 37.20 MiB/s [2024-11-26T18:23:39.354Z] 9798.80 IOPS, 38.28 MiB/s [2024-11-26T18:23:39.354Z] 10293.17 IOPS, 40.21 MiB/s [2024-11-26T18:23:39.354Z] 10654.43 IOPS, 41.62 MiB/s [2024-11-26T18:23:39.354Z] 10650.00 IOPS, 41.60 MiB/s [2024-11-26T18:23:39.354Z] 10527.56 IOPS, 41.12 MiB/s [2024-11-26T18:23:39.354Z] 10429.30 IOPS, 40.74 MiB/s [2024-11-26T18:23:39.354Z] 10356.09 IOPS, 40.45 MiB/s [2024-11-26T18:23:39.354Z] [2024-11-26 19:23:24.279213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:68240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:68248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:68256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:68264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:68272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:68280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:68288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:68296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.729 [2024-11-26 19:23:24.279378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:27:26.729 [2024-11-26 19:23:24.279388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:67480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.729 [2024-11-26 19:23:24.279394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.279404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:67488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.279415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.279426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:67496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.279431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.279441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:67504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.279446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.279457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:67512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.279462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.279472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:67520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.279477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.279488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:67528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.279493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.280892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:68304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.280901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.280913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:68312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.280918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.280929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:68320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.280934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.280945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:68328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.280950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.280961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:68336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.280967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.280978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:68344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.280983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.280994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:68352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.280998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:68360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:68368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:68376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:68384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:68392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:68400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:68408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:68416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:67536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:67544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:67552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:67560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:67568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:67576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:67584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:67592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:68424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.730 [2024-11-26 19:23:24.281315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:67600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:67608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:67616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:67624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:67632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:67640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:67648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:67656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:67664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:67672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:67680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:27:26.730 [2024-11-26 19:23:24.281516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:67688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.730 [2024-11-26 19:23:24.281521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:67696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:67704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:67712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:67720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:67728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:67736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:67744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:67752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:67760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:67768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:67776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:67784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:67792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:67800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:67808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:67816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:67824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:67832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:67840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:67848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:67856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:67864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:67872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:67880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:67888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:67896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:67904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.281985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.281997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:67912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.282002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:68432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:68440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:68448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:68456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:68464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:68472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:68480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:68488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.731 [2024-11-26 19:23:24.282534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:67920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.282553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:67928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.282572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:67936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.282591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:67944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.282610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:67952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.282631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:26.731 [2024-11-26 19:23:24.282645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:67960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.731 [2024-11-26 19:23:24.282650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:67968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:67976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282707] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:67984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:67992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:68000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:68008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:68016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:68024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:68032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:68040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:68048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:68056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:68064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:68072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.282985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:68080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.282992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:68088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:68096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:68104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:68112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:68120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:68128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:68136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:68144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:68152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:68160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:68168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283228] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:68176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:68184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:68192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:68200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:68208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:68216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:68224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:68232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.732 [2024-11-26 19:23:24.283374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:24.283389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:68496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:24.283395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:27:26.732 10260.25 IOPS, 40.08 MiB/s [2024-11-26T18:23:39.357Z] 9471.00 IOPS, 37.00 MiB/s [2024-11-26T18:23:39.357Z] 8794.50 IOPS, 34.35 MiB/s [2024-11-26T18:23:39.357Z] 8235.27 IOPS, 32.17 MiB/s [2024-11-26T18:23:39.357Z] 8524.31 IOPS, 33.30 MiB/s [2024-11-26T18:23:39.357Z] 8788.12 IOPS, 34.33 MiB/s [2024-11-26T18:23:39.357Z] 9220.06 IOPS, 36.02 MiB/s [2024-11-26T18:23:39.357Z] 9614.84 IOPS, 37.56 MiB/s [2024-11-26T18:23:39.357Z] 9876.70 IOPS, 38.58 MiB/s [2024-11-26T18:23:39.357Z] 10017.90 IOPS, 39.13 MiB/s [2024-11-26T18:23:39.357Z] 10146.64 IOPS, 39.64 MiB/s [2024-11-26T18:23:39.357Z] 10398.22 IOPS, 40.62 MiB/s [2024-11-26T18:23:39.357Z] 10663.58 IOPS, 41.65 MiB/s [2024-11-26T18:23:39.357Z] [2024-11-26 19:23:36.946089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:30904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:36.946124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:36.946154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:30920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:36.946161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:36.946171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:30936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:36.946182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:36.946193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:30952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:36.946198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:36.946208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:30968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:36.946213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:36.946224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:30984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:36.946229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:36.946240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:31000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.732 [2024-11-26 19:23:36.946245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:27:26.732 [2024-11-26 19:23:36.946256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:30152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:30184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:30216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:30744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:30776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:30800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:30832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:30872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:31024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.733 [2024-11-26 19:23:36.946499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:31040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.733 [2024-11-26 19:23:36.946515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:30288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:30312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:30336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:30368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:26.733 [2024-11-26 19:23:36.946662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:31056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.733 [2024-11-26 19:23:36.946788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:27:26.733 [2024-11-26 19:23:36.946799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:31072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:27:26.733 [2024-11-26 19:23:36.946804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:27:26.733 10788.44 IOPS, 42.14 MiB/s [2024-11-26T18:23:39.358Z] 10740.46 IOPS, 41.95 MiB/s [2024-11-26T18:23:39.358Z] Received shutdown signal, test time was about 26.769235 seconds 00:27:26.733 00:27:26.733 Latency(us) 00:27:26.733 [2024-11-26T18:23:39.358Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:26.733 Job: Nvme0n1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:27:26.733 Verification LBA range: start 0x0 length 0x4000 00:27:26.733 Nvme0n1 : 26.77 10707.05 41.82 0.00 0.00 11936.97 314.03 3019898.88 00:27:26.733 [2024-11-26T18:23:39.358Z] =================================================================================================================== 00:27:26.733 [2024-11-26T18:23:39.358Z] Total : 10707.05 41.82 0.00 0.00 11936.97 314.03 3019898.88 00:27:26.733 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@143 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@145 -- # trap - SIGINT SIGTERM EXIT 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@147 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@148 -- # nvmftestfini 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@516 -- # nvmfcleanup 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@121 -- # sync 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@124 -- # set +e 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:26.995 rmmod nvme_tcp 00:27:26.995 rmmod nvme_fabrics 00:27:26.995 rmmod nvme_keyring 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@128 -- # set -e 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@129 -- # return 0 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@517 -- # '[' -n 3998396 ']' 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@518 -- # killprocess 3998396 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # '[' -z 3998396 ']' 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@958 -- # kill -0 3998396 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # uname 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 3998396 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@972 -- # echo 'killing process with pid 3998396' 00:27:26.995 killing process with pid 3998396 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@973 -- # kill 3998396 00:27:26.995 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@978 -- # wait 3998396 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@297 -- # iptr 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@791 -- # iptables-save 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@791 -- # iptables-restore 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:27.256 19:23:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:29.169 19:23:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:29.169 00:27:29.169 real 0m41.654s 00:27:29.169 user 1m44.506s 00:27:29.169 sys 0m12.522s 00:27:29.169 19:23:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:29.169 19:23:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:27:29.169 ************************************ 00:27:29.169 END TEST nvmf_host_multipath_status 00:27:29.169 ************************************ 00:27:29.169 19:23:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@28 -- # run_test nvmf_discovery_remove_ifc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:27:29.169 19:23:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:27:29.169 19:23:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:29.169 19:23:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:29.430 ************************************ 00:27:29.430 START TEST nvmf_discovery_remove_ifc 00:27:29.430 ************************************ 00:27:29.430 19:23:41 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:27:29.430 * Looking for test storage... 00:27:29.430 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:29.430 19:23:41 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:27:29.430 19:23:41 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1693 -- # lcov --version 00:27:29.430 19:23:41 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # IFS=.-: 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # read -ra ver1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # IFS=.-: 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # read -ra ver2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@338 -- # local 'op=<' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@340 -- # ver1_l=2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@341 -- # ver2_l=1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@344 -- # case "$op" in 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@345 -- # : 1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # decimal 1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # ver1[v]=1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # decimal 2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # ver2[v]=2 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # return 0 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:27:29.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:29.430 --rc genhtml_branch_coverage=1 00:27:29.430 --rc genhtml_function_coverage=1 00:27:29.430 --rc genhtml_legend=1 00:27:29.430 --rc geninfo_all_blocks=1 00:27:29.430 --rc geninfo_unexecuted_blocks=1 00:27:29.430 00:27:29.430 ' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:27:29.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:29.430 --rc genhtml_branch_coverage=1 00:27:29.430 --rc genhtml_function_coverage=1 00:27:29.430 --rc genhtml_legend=1 00:27:29.430 --rc geninfo_all_blocks=1 00:27:29.430 --rc geninfo_unexecuted_blocks=1 00:27:29.430 00:27:29.430 ' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:27:29.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:29.430 --rc genhtml_branch_coverage=1 00:27:29.430 --rc genhtml_function_coverage=1 00:27:29.430 --rc genhtml_legend=1 00:27:29.430 --rc geninfo_all_blocks=1 00:27:29.430 --rc geninfo_unexecuted_blocks=1 00:27:29.430 00:27:29.430 ' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:27:29.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:29.430 --rc genhtml_branch_coverage=1 00:27:29.430 --rc genhtml_function_coverage=1 00:27:29.430 --rc genhtml_legend=1 00:27:29.430 --rc geninfo_all_blocks=1 00:27:29.430 --rc geninfo_unexecuted_blocks=1 00:27:29.430 00:27:29.430 ' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # uname -s 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:29.430 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@15 -- # shopt -s extglob 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@5 -- # export PATH 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@51 -- # : 0 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:29.691 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@14 -- # '[' tcp == rdma ']' 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@19 -- # discovery_port=8009 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@20 -- # discovery_nqn=nqn.2014-08.org.nvmexpress.discovery 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@23 -- # nqn=nqn.2016-06.io.spdk:cnode 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@25 -- # host_nqn=nqn.2021-12.io.spdk:test 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@26 -- # host_sock=/tmp/host.sock 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@39 -- # nvmftestinit 00:27:29.691 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@476 -- # prepare_net_devs 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@438 -- # local -g is_hw=no 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # remove_spdk_ns 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@309 -- # xtrace_disable 00:27:29.692 19:23:42 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # pci_devs=() 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # net_devs=() 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # e810=() 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # local -ga e810 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # x722=() 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # local -ga x722 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # mlx=() 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # local -ga mlx 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:27:37.858 Found 0000:31:00.0 (0x8086 - 0x159b) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:27:37.858 Found 0000:31:00.1 (0x8086 - 0x159b) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:37.858 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:27:37.859 Found net devices under 0000:31:00.0: cvl_0_0 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:27:37.859 Found net devices under 0000:31:00.1: cvl_0_1 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # is_hw=yes 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:37.859 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:37.859 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.522 ms 00:27:37.859 00:27:37.859 --- 10.0.0.2 ping statistics --- 00:27:37.859 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:37.859 rtt min/avg/max/mdev = 0.522/0.522/0.522/0.000 ms 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:37.859 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:37.859 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.305 ms 00:27:37.859 00:27:37.859 --- 10.0.0.1 ping statistics --- 00:27:37.859 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:37.859 rtt min/avg/max/mdev = 0.305/0.305/0.305/0.000 ms 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@450 -- # return 0 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@40 -- # nvmfappstart -m 0x2 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@726 -- # xtrace_disable 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@509 -- # nvmfpid=4009293 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@510 -- # waitforlisten 4009293 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # '[' -z 4009293 ']' 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:37.859 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:37.859 19:23:50 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:38.121 [2024-11-26 19:23:50.524008] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:27:38.121 [2024-11-26 19:23:50.524067] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:38.121 [2024-11-26 19:23:50.631481] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:38.121 [2024-11-26 19:23:50.681355] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:38.121 [2024-11-26 19:23:50.681412] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:38.121 [2024-11-26 19:23:50.681421] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:38.121 [2024-11-26 19:23:50.681429] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:38.121 [2024-11-26 19:23:50.681435] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:38.121 [2024-11-26 19:23:50.682277] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@868 -- # return 0 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@732 -- # xtrace_disable 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@43 -- # rpc_cmd 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:39.064 [2024-11-26 19:23:51.397239] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:39.064 [2024-11-26 19:23:51.405477] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:27:39.064 null0 00:27:39.064 [2024-11-26 19:23:51.437429] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@59 -- # hostpid=4009566 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock --wait-for-rpc -L bdev_nvme 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@60 -- # waitforlisten 4009566 /tmp/host.sock 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # '[' -z 4009566 ']' 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@839 -- # local rpc_addr=/tmp/host.sock 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:27:39.064 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:39.064 19:23:51 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:39.064 [2024-11-26 19:23:51.523700] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:27:39.064 [2024-11-26 19:23:51.523762] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4009566 ] 00:27:39.064 [2024-11-26 19:23:51.606534] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:39.064 [2024-11-26 19:23:51.648676] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@868 -- # return 0 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@62 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@65 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_set_options -e 1 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@66 -- # rpc_cmd -s /tmp/host.sock framework_start_init 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@69 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test --ctrlr-loss-timeout-sec 2 --reconnect-delay-sec 1 --fast-io-fail-timeout-sec 1 --wait-for-attach 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.007 19:23:52 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:40.948 [2024-11-26 19:23:53.447019] bdev_nvme.c:7484:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:27:40.948 [2024-11-26 19:23:53.447042] bdev_nvme.c:7570:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:27:40.948 [2024-11-26 19:23:53.447056] bdev_nvme.c:7447:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:27:41.209 [2024-11-26 19:23:53.574461] bdev_nvme.c:7413:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:27:41.209 [2024-11-26 19:23:53.635233] bdev_nvme.c:5636:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr was created to 10.0.0.2:4420 00:27:41.209 [2024-11-26 19:23:53.636396] bdev_nvme.c:1985:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x146d690:1 started. 00:27:41.209 [2024-11-26 19:23:53.637966] bdev_nvme.c:8280:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:27:41.209 [2024-11-26 19:23:53.638011] bdev_nvme.c:8280:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:27:41.209 [2024-11-26 19:23:53.638033] bdev_nvme.c:8280:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:27:41.209 [2024-11-26 19:23:53.638047] bdev_nvme.c:7303:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:27:41.209 [2024-11-26 19:23:53.638068] bdev_nvme.c:7262:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@72 -- # wait_for_bdev nvme0n1 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:41.209 [2024-11-26 19:23:53.645285] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x146d690 was disconnected and freed. delete nvme_qpair. 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:41.209 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != \n\v\m\e\0\n\1 ]] 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@75 -- # ip netns exec cvl_0_0_ns_spdk ip addr del 10.0.0.2/24 dev cvl_0_0 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@76 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 down 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@79 -- # wait_for_bdev '' 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:41.210 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:41.470 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.470 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:27:41.470 19:23:53 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:27:42.411 19:23:54 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:43.351 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.611 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:27:43.611 19:23:55 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:44.552 19:23:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:44.552 19:23:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:44.552 19:23:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:44.552 19:23:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:44.552 19:23:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:44.552 19:23:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:44.552 19:23:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:44.552 19:23:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.552 19:23:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:27:44.552 19:23:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:27:45.493 19:23:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:46.874 [2024-11-26 19:23:59.078568] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 110: Connection timed out 00:27:46.874 [2024-11-26 19:23:59.078609] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:27:46.874 [2024-11-26 19:23:59.078622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:46.874 [2024-11-26 19:23:59.078632] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:27:46.874 [2024-11-26 19:23:59.078639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:46.874 [2024-11-26 19:23:59.078648] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:27:46.874 [2024-11-26 19:23:59.078655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:46.874 [2024-11-26 19:23:59.078663] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:27:46.874 [2024-11-26 19:23:59.078671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:46.874 [2024-11-26 19:23:59.078679] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:27:46.874 [2024-11-26 19:23:59.078687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:46.874 [2024-11-26 19:23:59.078695] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x144a0a0 is same with the state(6) to be set 00:27:46.874 [2024-11-26 19:23:59.088590] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x144a0a0 (9): Bad file descriptor 00:27:46.874 [2024-11-26 19:23:59.098625] bdev_nvme.c:2545:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:27:46.874 [2024-11-26 19:23:59.098639] bdev_nvme.c:2533:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:27:46.874 [2024-11-26 19:23:59.098644] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:27:46.874 [2024-11-26 19:23:59.098650] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:27:46.874 [2024-11-26 19:23:59.098671] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:27:46.874 19:23:59 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:46.874 19:23:59 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:46.874 19:23:59 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:46.874 19:23:59 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:46.874 19:23:59 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:46.874 19:23:59 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:46.874 19:23:59 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:47.814 [2024-11-26 19:24:00.155909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 110 00:27:47.814 [2024-11-26 19:24:00.155960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x144a0a0 with addr=10.0.0.2, port=4420 00:27:47.814 [2024-11-26 19:24:00.155980] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x144a0a0 is same with the state(6) to be set 00:27:47.814 [2024-11-26 19:24:00.156012] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x144a0a0 (9): Bad file descriptor 00:27:47.814 [2024-11-26 19:24:00.156416] bdev_nvme.c:3168:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] Unable to perform failover, already in progress. 00:27:47.814 [2024-11-26 19:24:00.156443] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:27:47.814 [2024-11-26 19:24:00.156451] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:27:47.814 [2024-11-26 19:24:00.156460] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:27:47.814 [2024-11-26 19:24:00.156468] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:27:47.814 [2024-11-26 19:24:00.156475] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:27:47.814 [2024-11-26 19:24:00.156480] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:27:47.814 [2024-11-26 19:24:00.156488] bdev_nvme.c:2129:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:27:47.814 [2024-11-26 19:24:00.156493] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:27:47.814 19:24:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.814 19:24:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:27:47.814 19:24:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:48.757 [2024-11-26 19:24:01.158873] bdev_nvme.c:2517:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:27:48.757 [2024-11-26 19:24:01.158897] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:27:48.757 [2024-11-26 19:24:01.158910] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:27:48.757 [2024-11-26 19:24:01.158919] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:27:48.757 [2024-11-26 19:24:01.158927] nvme_ctrlr.c:1098:nvme_ctrlr_fail: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] already in failed state 00:27:48.757 [2024-11-26 19:24:01.158935] bdev_nvme.c:2507:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:27:48.757 [2024-11-26 19:24:01.158941] bdev_nvme.c:2274:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:27:48.757 [2024-11-26 19:24:01.158946] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:27:48.757 [2024-11-26 19:24:01.158971] bdev_nvme.c:7235:remove_discovery_entry: *INFO*: Discovery[10.0.0.2:8009] Remove discovery entry: nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 00:27:48.757 [2024-11-26 19:24:01.158996] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:27:48.757 [2024-11-26 19:24:01.159008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:48.757 [2024-11-26 19:24:01.159019] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:27:48.757 [2024-11-26 19:24:01.159027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:48.757 [2024-11-26 19:24:01.159036] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:27:48.757 [2024-11-26 19:24:01.159044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:48.757 [2024-11-26 19:24:01.159056] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:27:48.757 [2024-11-26 19:24:01.159064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:48.757 [2024-11-26 19:24:01.159073] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:27:48.757 [2024-11-26 19:24:01.159080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:48.757 [2024-11-26 19:24:01.159088] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery, 1] in failed state. 00:27:48.757 [2024-11-26 19:24:01.159356] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1439390 (9): Bad file descriptor 00:27:48.757 [2024-11-26 19:24:01.160369] nvme_fabric.c: 214:nvme_fabric_prop_get_cmd_async: *ERROR*: Failed to send Property Get fabrics command 00:27:48.757 [2024-11-26 19:24:01.160380] nvme_ctrlr.c:1217:nvme_ctrlr_shutdown_async: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery, 1] Failed to read the CC register 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != '' ]] 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@82 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@83 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@86 -- # wait_for_bdev nvme1n1 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:48.757 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.018 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:27:49.018 19:24:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:27:49.961 19:24:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:50.904 [2024-11-26 19:24:03.210684] bdev_nvme.c:7484:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:27:50.904 [2024-11-26 19:24:03.210704] bdev_nvme.c:7570:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:27:50.904 [2024-11-26 19:24:03.210718] bdev_nvme.c:7447:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:27:50.904 [2024-11-26 19:24:03.339116] bdev_nvme.c:7413:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme1 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:27:50.904 19:24:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:27:50.904 [2024-11-26 19:24:03.520275] bdev_nvme.c:5636:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] ctrlr was created to 10.0.0.2:4420 00:27:50.904 [2024-11-26 19:24:03.521288] bdev_nvme.c:1985:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] Connecting qpair 0x144c3e0:1 started. 00:27:50.904 [2024-11-26 19:24:03.522524] bdev_nvme.c:8280:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:27:50.904 [2024-11-26 19:24:03.522560] bdev_nvme.c:8280:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:27:50.904 [2024-11-26 19:24:03.522580] bdev_nvme.c:8280:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:27:50.904 [2024-11-26 19:24:03.522594] bdev_nvme.c:7303:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme1 done 00:27:50.904 [2024-11-26 19:24:03.522603] bdev_nvme.c:7262:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:27:51.164 [2024-11-26 19:24:03.529373] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] qpair 0x144c3e0 was disconnected and freed. delete nvme_qpair. 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme1n1 != \n\v\m\e\1\n\1 ]] 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@88 -- # trap - SIGINT SIGTERM EXIT 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@90 -- # killprocess 4009566 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # '[' -z 4009566 ']' 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@958 -- # kill -0 4009566 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # uname 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4009566 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4009566' 00:27:52.108 killing process with pid 4009566 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@973 -- # kill 4009566 00:27:52.108 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@978 -- # wait 4009566 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@91 -- # nvmftestfini 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@516 -- # nvmfcleanup 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@121 -- # sync 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@124 -- # set +e 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:52.369 rmmod nvme_tcp 00:27:52.369 rmmod nvme_fabrics 00:27:52.369 rmmod nvme_keyring 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@128 -- # set -e 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@129 -- # return 0 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@517 -- # '[' -n 4009293 ']' 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@518 -- # killprocess 4009293 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # '[' -z 4009293 ']' 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@958 -- # kill -0 4009293 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # uname 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4009293 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4009293' 00:27:52.369 killing process with pid 4009293 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@973 -- # kill 4009293 00:27:52.369 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@978 -- # wait 4009293 00:27:52.630 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:27:52.630 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:27:52.630 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:27:52.630 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@297 -- # iptr 00:27:52.630 19:24:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:27:52.630 19:24:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@791 -- # iptables-restore 00:27:52.630 19:24:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@791 -- # iptables-save 00:27:52.630 19:24:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:52.630 19:24:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:52.630 19:24:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:52.630 19:24:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:52.630 19:24:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:54.543 19:24:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:54.543 00:27:54.543 real 0m25.263s 00:27:54.543 user 0m29.541s 00:27:54.543 sys 0m7.809s 00:27:54.543 19:24:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:54.543 19:24:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:27:54.543 ************************************ 00:27:54.543 END TEST nvmf_discovery_remove_ifc 00:27:54.543 ************************************ 00:27:54.543 19:24:07 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@29 -- # run_test nvmf_identify_kernel_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:27:54.543 19:24:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:27:54.543 19:24:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:54.543 19:24:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:54.543 ************************************ 00:27:54.543 START TEST nvmf_identify_kernel_target 00:27:54.543 ************************************ 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:27:54.803 * Looking for test storage... 00:27:54.803 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1693 -- # lcov --version 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # IFS=.-: 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # read -ra ver1 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # IFS=.-: 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # read -ra ver2 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@338 -- # local 'op=<' 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@340 -- # ver1_l=2 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@341 -- # ver2_l=1 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@344 -- # case "$op" in 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@345 -- # : 1 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # decimal 1 00:27:54.803 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=1 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 1 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # ver1[v]=1 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # decimal 2 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=2 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 2 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # ver2[v]=2 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # return 0 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:27:54.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:54.804 --rc genhtml_branch_coverage=1 00:27:54.804 --rc genhtml_function_coverage=1 00:27:54.804 --rc genhtml_legend=1 00:27:54.804 --rc geninfo_all_blocks=1 00:27:54.804 --rc geninfo_unexecuted_blocks=1 00:27:54.804 00:27:54.804 ' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:27:54.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:54.804 --rc genhtml_branch_coverage=1 00:27:54.804 --rc genhtml_function_coverage=1 00:27:54.804 --rc genhtml_legend=1 00:27:54.804 --rc geninfo_all_blocks=1 00:27:54.804 --rc geninfo_unexecuted_blocks=1 00:27:54.804 00:27:54.804 ' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:27:54.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:54.804 --rc genhtml_branch_coverage=1 00:27:54.804 --rc genhtml_function_coverage=1 00:27:54.804 --rc genhtml_legend=1 00:27:54.804 --rc geninfo_all_blocks=1 00:27:54.804 --rc geninfo_unexecuted_blocks=1 00:27:54.804 00:27:54.804 ' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:27:54.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:54.804 --rc genhtml_branch_coverage=1 00:27:54.804 --rc genhtml_function_coverage=1 00:27:54.804 --rc genhtml_legend=1 00:27:54.804 --rc geninfo_all_blocks=1 00:27:54.804 --rc geninfo_unexecuted_blocks=1 00:27:54.804 00:27:54.804 ' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # uname -s 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@15 -- # shopt -s extglob 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@5 -- # export PATH 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@51 -- # : 0 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:54.804 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@11 -- # nvmftestinit 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:27:54.804 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:27:54.805 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:54.805 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:54.805 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:54.805 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:27:54.805 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:27:54.805 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@309 -- # xtrace_disable 00:27:54.805 19:24:07 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # pci_devs=() 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # net_devs=() 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # e810=() 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # local -ga e810 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # x722=() 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # local -ga x722 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # mlx=() 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # local -ga mlx 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:28:03.076 Found 0000:31:00.0 (0x8086 - 0x159b) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:28:03.076 Found 0000:31:00.1 (0x8086 - 0x159b) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:28:03.076 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:28:03.077 Found net devices under 0000:31:00.0: cvl_0_0 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:28:03.077 Found net devices under 0000:31:00.1: cvl_0_1 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # is_hw=yes 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:28:03.077 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:28:03.366 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:28:03.366 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.709 ms 00:28:03.366 00:28:03.366 --- 10.0.0.2 ping statistics --- 00:28:03.366 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:03.366 rtt min/avg/max/mdev = 0.709/0.709/0.709/0.000 ms 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:28:03.366 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:28:03.366 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.255 ms 00:28:03.366 00:28:03.366 --- 10.0.0.1 ping statistics --- 00:28:03.366 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:03.366 rtt min/avg/max/mdev = 0.255/0.255/0.255/0.000 ms 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@450 -- # return 0 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:28:03.366 19:24:15 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@13 -- # trap 'nvmftestfini || :; clean_kernel_target' EXIT 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # get_main_ns_ip 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@769 -- # local ip 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # target_ip=10.0.0.1 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@16 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@660 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@662 -- # nvmet=/sys/kernel/config/nvmet 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@663 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@664 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@665 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@667 -- # local block nvme 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@669 -- # [[ ! -e /sys/module/nvmet ]] 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@670 -- # modprobe nvmet 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet ]] 00:28:03.627 19:24:16 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@675 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:28:07.837 Waiting for block devices as requested 00:28:07.837 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:28:07.837 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:28:08.099 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:28:08.099 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:28:08.361 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:28:08.361 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:28:08.361 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:28:08.622 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:28:08.622 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:28:08.622 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@678 -- # for block in /sys/block/nvme* 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@679 -- # [[ -e /sys/block/nvme0n1 ]] 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@680 -- # is_block_zoned nvme0n1 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@681 -- # block_in_use nvme0n1 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:28:08.883 No valid GPT data, bailing 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # pt= 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@395 -- # return 1 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@681 -- # nvme=/dev/nvme0n1 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@684 -- # [[ -b /dev/nvme0n1 ]] 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:28:08.883 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@687 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@688 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@693 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@695 -- # echo 1 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@696 -- # echo /dev/nvme0n1 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@697 -- # echo 1 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@699 -- # echo 10.0.0.1 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@700 -- # echo tcp 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@701 -- # echo 4420 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@702 -- # echo ipv4 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@705 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@708 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -a 10.0.0.1 -t tcp -s 4420 00:28:09.146 00:28:09.146 Discovery Log Number of Records 2, Generation counter 2 00:28:09.146 =====Discovery Log Entry 0====== 00:28:09.146 trtype: tcp 00:28:09.146 adrfam: ipv4 00:28:09.146 subtype: current discovery subsystem 00:28:09.146 treq: not specified, sq flow control disable supported 00:28:09.146 portid: 1 00:28:09.146 trsvcid: 4420 00:28:09.146 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:28:09.146 traddr: 10.0.0.1 00:28:09.146 eflags: none 00:28:09.146 sectype: none 00:28:09.146 =====Discovery Log Entry 1====== 00:28:09.146 trtype: tcp 00:28:09.146 adrfam: ipv4 00:28:09.146 subtype: nvme subsystem 00:28:09.146 treq: not specified, sq flow control disable supported 00:28:09.146 portid: 1 00:28:09.146 trsvcid: 4420 00:28:09.146 subnqn: nqn.2016-06.io.spdk:testnqn 00:28:09.146 traddr: 10.0.0.1 00:28:09.146 eflags: none 00:28:09.146 sectype: none 00:28:09.146 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 00:28:09.146 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' 00:28:09.146 ===================================================== 00:28:09.146 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2014-08.org.nvmexpress.discovery 00:28:09.146 ===================================================== 00:28:09.146 Controller Capabilities/Features 00:28:09.146 ================================ 00:28:09.146 Vendor ID: 0000 00:28:09.146 Subsystem Vendor ID: 0000 00:28:09.146 Serial Number: 3998a90590ac99feb9f2 00:28:09.146 Model Number: Linux 00:28:09.146 Firmware Version: 6.8.9-20 00:28:09.146 Recommended Arb Burst: 0 00:28:09.146 IEEE OUI Identifier: 00 00 00 00:28:09.146 Multi-path I/O 00:28:09.146 May have multiple subsystem ports: No 00:28:09.146 May have multiple controllers: No 00:28:09.146 Associated with SR-IOV VF: No 00:28:09.146 Max Data Transfer Size: Unlimited 00:28:09.146 Max Number of Namespaces: 0 00:28:09.146 Max Number of I/O Queues: 1024 00:28:09.146 NVMe Specification Version (VS): 1.3 00:28:09.146 NVMe Specification Version (Identify): 1.3 00:28:09.146 Maximum Queue Entries: 1024 00:28:09.146 Contiguous Queues Required: No 00:28:09.146 Arbitration Mechanisms Supported 00:28:09.146 Weighted Round Robin: Not Supported 00:28:09.146 Vendor Specific: Not Supported 00:28:09.146 Reset Timeout: 7500 ms 00:28:09.146 Doorbell Stride: 4 bytes 00:28:09.146 NVM Subsystem Reset: Not Supported 00:28:09.146 Command Sets Supported 00:28:09.146 NVM Command Set: Supported 00:28:09.146 Boot Partition: Not Supported 00:28:09.146 Memory Page Size Minimum: 4096 bytes 00:28:09.146 Memory Page Size Maximum: 4096 bytes 00:28:09.146 Persistent Memory Region: Not Supported 00:28:09.146 Optional Asynchronous Events Supported 00:28:09.146 Namespace Attribute Notices: Not Supported 00:28:09.146 Firmware Activation Notices: Not Supported 00:28:09.146 ANA Change Notices: Not Supported 00:28:09.146 PLE Aggregate Log Change Notices: Not Supported 00:28:09.146 LBA Status Info Alert Notices: Not Supported 00:28:09.146 EGE Aggregate Log Change Notices: Not Supported 00:28:09.146 Normal NVM Subsystem Shutdown event: Not Supported 00:28:09.146 Zone Descriptor Change Notices: Not Supported 00:28:09.146 Discovery Log Change Notices: Supported 00:28:09.146 Controller Attributes 00:28:09.146 128-bit Host Identifier: Not Supported 00:28:09.146 Non-Operational Permissive Mode: Not Supported 00:28:09.146 NVM Sets: Not Supported 00:28:09.146 Read Recovery Levels: Not Supported 00:28:09.146 Endurance Groups: Not Supported 00:28:09.146 Predictable Latency Mode: Not Supported 00:28:09.146 Traffic Based Keep ALive: Not Supported 00:28:09.146 Namespace Granularity: Not Supported 00:28:09.146 SQ Associations: Not Supported 00:28:09.146 UUID List: Not Supported 00:28:09.146 Multi-Domain Subsystem: Not Supported 00:28:09.146 Fixed Capacity Management: Not Supported 00:28:09.146 Variable Capacity Management: Not Supported 00:28:09.146 Delete Endurance Group: Not Supported 00:28:09.146 Delete NVM Set: Not Supported 00:28:09.146 Extended LBA Formats Supported: Not Supported 00:28:09.146 Flexible Data Placement Supported: Not Supported 00:28:09.146 00:28:09.146 Controller Memory Buffer Support 00:28:09.146 ================================ 00:28:09.146 Supported: No 00:28:09.146 00:28:09.146 Persistent Memory Region Support 00:28:09.146 ================================ 00:28:09.146 Supported: No 00:28:09.146 00:28:09.146 Admin Command Set Attributes 00:28:09.146 ============================ 00:28:09.146 Security Send/Receive: Not Supported 00:28:09.146 Format NVM: Not Supported 00:28:09.146 Firmware Activate/Download: Not Supported 00:28:09.146 Namespace Management: Not Supported 00:28:09.146 Device Self-Test: Not Supported 00:28:09.146 Directives: Not Supported 00:28:09.146 NVMe-MI: Not Supported 00:28:09.146 Virtualization Management: Not Supported 00:28:09.146 Doorbell Buffer Config: Not Supported 00:28:09.146 Get LBA Status Capability: Not Supported 00:28:09.146 Command & Feature Lockdown Capability: Not Supported 00:28:09.146 Abort Command Limit: 1 00:28:09.146 Async Event Request Limit: 1 00:28:09.146 Number of Firmware Slots: N/A 00:28:09.146 Firmware Slot 1 Read-Only: N/A 00:28:09.146 Firmware Activation Without Reset: N/A 00:28:09.146 Multiple Update Detection Support: N/A 00:28:09.146 Firmware Update Granularity: No Information Provided 00:28:09.146 Per-Namespace SMART Log: No 00:28:09.146 Asymmetric Namespace Access Log Page: Not Supported 00:28:09.147 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:28:09.147 Command Effects Log Page: Not Supported 00:28:09.147 Get Log Page Extended Data: Supported 00:28:09.147 Telemetry Log Pages: Not Supported 00:28:09.147 Persistent Event Log Pages: Not Supported 00:28:09.147 Supported Log Pages Log Page: May Support 00:28:09.147 Commands Supported & Effects Log Page: Not Supported 00:28:09.147 Feature Identifiers & Effects Log Page:May Support 00:28:09.147 NVMe-MI Commands & Effects Log Page: May Support 00:28:09.147 Data Area 4 for Telemetry Log: Not Supported 00:28:09.147 Error Log Page Entries Supported: 1 00:28:09.147 Keep Alive: Not Supported 00:28:09.147 00:28:09.147 NVM Command Set Attributes 00:28:09.147 ========================== 00:28:09.147 Submission Queue Entry Size 00:28:09.147 Max: 1 00:28:09.147 Min: 1 00:28:09.147 Completion Queue Entry Size 00:28:09.147 Max: 1 00:28:09.147 Min: 1 00:28:09.147 Number of Namespaces: 0 00:28:09.147 Compare Command: Not Supported 00:28:09.147 Write Uncorrectable Command: Not Supported 00:28:09.147 Dataset Management Command: Not Supported 00:28:09.147 Write Zeroes Command: Not Supported 00:28:09.147 Set Features Save Field: Not Supported 00:28:09.147 Reservations: Not Supported 00:28:09.147 Timestamp: Not Supported 00:28:09.147 Copy: Not Supported 00:28:09.147 Volatile Write Cache: Not Present 00:28:09.147 Atomic Write Unit (Normal): 1 00:28:09.147 Atomic Write Unit (PFail): 1 00:28:09.147 Atomic Compare & Write Unit: 1 00:28:09.147 Fused Compare & Write: Not Supported 00:28:09.147 Scatter-Gather List 00:28:09.147 SGL Command Set: Supported 00:28:09.147 SGL Keyed: Not Supported 00:28:09.147 SGL Bit Bucket Descriptor: Not Supported 00:28:09.147 SGL Metadata Pointer: Not Supported 00:28:09.147 Oversized SGL: Not Supported 00:28:09.147 SGL Metadata Address: Not Supported 00:28:09.147 SGL Offset: Supported 00:28:09.147 Transport SGL Data Block: Not Supported 00:28:09.147 Replay Protected Memory Block: Not Supported 00:28:09.147 00:28:09.147 Firmware Slot Information 00:28:09.147 ========================= 00:28:09.147 Active slot: 0 00:28:09.147 00:28:09.147 00:28:09.147 Error Log 00:28:09.147 ========= 00:28:09.147 00:28:09.147 Active Namespaces 00:28:09.147 ================= 00:28:09.147 Discovery Log Page 00:28:09.147 ================== 00:28:09.147 Generation Counter: 2 00:28:09.147 Number of Records: 2 00:28:09.147 Record Format: 0 00:28:09.147 00:28:09.147 Discovery Log Entry 0 00:28:09.147 ---------------------- 00:28:09.147 Transport Type: 3 (TCP) 00:28:09.147 Address Family: 1 (IPv4) 00:28:09.147 Subsystem Type: 3 (Current Discovery Subsystem) 00:28:09.147 Entry Flags: 00:28:09.147 Duplicate Returned Information: 0 00:28:09.147 Explicit Persistent Connection Support for Discovery: 0 00:28:09.147 Transport Requirements: 00:28:09.147 Secure Channel: Not Specified 00:28:09.147 Port ID: 1 (0x0001) 00:28:09.147 Controller ID: 65535 (0xffff) 00:28:09.147 Admin Max SQ Size: 32 00:28:09.147 Transport Service Identifier: 4420 00:28:09.147 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:28:09.147 Transport Address: 10.0.0.1 00:28:09.147 Discovery Log Entry 1 00:28:09.147 ---------------------- 00:28:09.147 Transport Type: 3 (TCP) 00:28:09.147 Address Family: 1 (IPv4) 00:28:09.147 Subsystem Type: 2 (NVM Subsystem) 00:28:09.147 Entry Flags: 00:28:09.147 Duplicate Returned Information: 0 00:28:09.147 Explicit Persistent Connection Support for Discovery: 0 00:28:09.147 Transport Requirements: 00:28:09.147 Secure Channel: Not Specified 00:28:09.147 Port ID: 1 (0x0001) 00:28:09.147 Controller ID: 65535 (0xffff) 00:28:09.147 Admin Max SQ Size: 32 00:28:09.147 Transport Service Identifier: 4420 00:28:09.147 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:testnqn 00:28:09.147 Transport Address: 10.0.0.1 00:28:09.147 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:28:09.409 get_feature(0x01) failed 00:28:09.409 get_feature(0x02) failed 00:28:09.409 get_feature(0x04) failed 00:28:09.409 ===================================================== 00:28:09.409 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:28:09.409 ===================================================== 00:28:09.409 Controller Capabilities/Features 00:28:09.409 ================================ 00:28:09.409 Vendor ID: 0000 00:28:09.409 Subsystem Vendor ID: 0000 00:28:09.409 Serial Number: 9fd78a9d2688ab30b44e 00:28:09.409 Model Number: SPDK-nqn.2016-06.io.spdk:testnqn 00:28:09.409 Firmware Version: 6.8.9-20 00:28:09.409 Recommended Arb Burst: 6 00:28:09.409 IEEE OUI Identifier: 00 00 00 00:28:09.409 Multi-path I/O 00:28:09.409 May have multiple subsystem ports: Yes 00:28:09.409 May have multiple controllers: Yes 00:28:09.409 Associated with SR-IOV VF: No 00:28:09.409 Max Data Transfer Size: Unlimited 00:28:09.409 Max Number of Namespaces: 1024 00:28:09.409 Max Number of I/O Queues: 128 00:28:09.409 NVMe Specification Version (VS): 1.3 00:28:09.409 NVMe Specification Version (Identify): 1.3 00:28:09.409 Maximum Queue Entries: 1024 00:28:09.409 Contiguous Queues Required: No 00:28:09.409 Arbitration Mechanisms Supported 00:28:09.409 Weighted Round Robin: Not Supported 00:28:09.409 Vendor Specific: Not Supported 00:28:09.409 Reset Timeout: 7500 ms 00:28:09.409 Doorbell Stride: 4 bytes 00:28:09.409 NVM Subsystem Reset: Not Supported 00:28:09.409 Command Sets Supported 00:28:09.409 NVM Command Set: Supported 00:28:09.409 Boot Partition: Not Supported 00:28:09.409 Memory Page Size Minimum: 4096 bytes 00:28:09.409 Memory Page Size Maximum: 4096 bytes 00:28:09.409 Persistent Memory Region: Not Supported 00:28:09.409 Optional Asynchronous Events Supported 00:28:09.409 Namespace Attribute Notices: Supported 00:28:09.409 Firmware Activation Notices: Not Supported 00:28:09.409 ANA Change Notices: Supported 00:28:09.409 PLE Aggregate Log Change Notices: Not Supported 00:28:09.409 LBA Status Info Alert Notices: Not Supported 00:28:09.409 EGE Aggregate Log Change Notices: Not Supported 00:28:09.409 Normal NVM Subsystem Shutdown event: Not Supported 00:28:09.409 Zone Descriptor Change Notices: Not Supported 00:28:09.409 Discovery Log Change Notices: Not Supported 00:28:09.409 Controller Attributes 00:28:09.409 128-bit Host Identifier: Supported 00:28:09.409 Non-Operational Permissive Mode: Not Supported 00:28:09.409 NVM Sets: Not Supported 00:28:09.409 Read Recovery Levels: Not Supported 00:28:09.409 Endurance Groups: Not Supported 00:28:09.409 Predictable Latency Mode: Not Supported 00:28:09.409 Traffic Based Keep ALive: Supported 00:28:09.409 Namespace Granularity: Not Supported 00:28:09.409 SQ Associations: Not Supported 00:28:09.409 UUID List: Not Supported 00:28:09.409 Multi-Domain Subsystem: Not Supported 00:28:09.409 Fixed Capacity Management: Not Supported 00:28:09.409 Variable Capacity Management: Not Supported 00:28:09.409 Delete Endurance Group: Not Supported 00:28:09.409 Delete NVM Set: Not Supported 00:28:09.409 Extended LBA Formats Supported: Not Supported 00:28:09.409 Flexible Data Placement Supported: Not Supported 00:28:09.409 00:28:09.409 Controller Memory Buffer Support 00:28:09.409 ================================ 00:28:09.409 Supported: No 00:28:09.409 00:28:09.409 Persistent Memory Region Support 00:28:09.409 ================================ 00:28:09.409 Supported: No 00:28:09.409 00:28:09.409 Admin Command Set Attributes 00:28:09.409 ============================ 00:28:09.410 Security Send/Receive: Not Supported 00:28:09.410 Format NVM: Not Supported 00:28:09.410 Firmware Activate/Download: Not Supported 00:28:09.410 Namespace Management: Not Supported 00:28:09.410 Device Self-Test: Not Supported 00:28:09.410 Directives: Not Supported 00:28:09.410 NVMe-MI: Not Supported 00:28:09.410 Virtualization Management: Not Supported 00:28:09.410 Doorbell Buffer Config: Not Supported 00:28:09.410 Get LBA Status Capability: Not Supported 00:28:09.410 Command & Feature Lockdown Capability: Not Supported 00:28:09.410 Abort Command Limit: 4 00:28:09.410 Async Event Request Limit: 4 00:28:09.410 Number of Firmware Slots: N/A 00:28:09.410 Firmware Slot 1 Read-Only: N/A 00:28:09.410 Firmware Activation Without Reset: N/A 00:28:09.410 Multiple Update Detection Support: N/A 00:28:09.410 Firmware Update Granularity: No Information Provided 00:28:09.410 Per-Namespace SMART Log: Yes 00:28:09.410 Asymmetric Namespace Access Log Page: Supported 00:28:09.410 ANA Transition Time : 10 sec 00:28:09.410 00:28:09.410 Asymmetric Namespace Access Capabilities 00:28:09.410 ANA Optimized State : Supported 00:28:09.410 ANA Non-Optimized State : Supported 00:28:09.410 ANA Inaccessible State : Supported 00:28:09.410 ANA Persistent Loss State : Supported 00:28:09.410 ANA Change State : Supported 00:28:09.410 ANAGRPID is not changed : No 00:28:09.410 Non-Zero ANAGRPID for NS Mgmt Cmd : Not Supported 00:28:09.410 00:28:09.410 ANA Group Identifier Maximum : 128 00:28:09.410 Number of ANA Group Identifiers : 128 00:28:09.410 Max Number of Allowed Namespaces : 1024 00:28:09.410 Subsystem NQN: nqn.2016-06.io.spdk:testnqn 00:28:09.410 Command Effects Log Page: Supported 00:28:09.410 Get Log Page Extended Data: Supported 00:28:09.410 Telemetry Log Pages: Not Supported 00:28:09.410 Persistent Event Log Pages: Not Supported 00:28:09.410 Supported Log Pages Log Page: May Support 00:28:09.410 Commands Supported & Effects Log Page: Not Supported 00:28:09.410 Feature Identifiers & Effects Log Page:May Support 00:28:09.410 NVMe-MI Commands & Effects Log Page: May Support 00:28:09.410 Data Area 4 for Telemetry Log: Not Supported 00:28:09.410 Error Log Page Entries Supported: 128 00:28:09.410 Keep Alive: Supported 00:28:09.410 Keep Alive Granularity: 1000 ms 00:28:09.410 00:28:09.410 NVM Command Set Attributes 00:28:09.410 ========================== 00:28:09.410 Submission Queue Entry Size 00:28:09.410 Max: 64 00:28:09.410 Min: 64 00:28:09.410 Completion Queue Entry Size 00:28:09.410 Max: 16 00:28:09.410 Min: 16 00:28:09.410 Number of Namespaces: 1024 00:28:09.410 Compare Command: Not Supported 00:28:09.410 Write Uncorrectable Command: Not Supported 00:28:09.410 Dataset Management Command: Supported 00:28:09.410 Write Zeroes Command: Supported 00:28:09.410 Set Features Save Field: Not Supported 00:28:09.410 Reservations: Not Supported 00:28:09.410 Timestamp: Not Supported 00:28:09.410 Copy: Not Supported 00:28:09.410 Volatile Write Cache: Present 00:28:09.410 Atomic Write Unit (Normal): 1 00:28:09.410 Atomic Write Unit (PFail): 1 00:28:09.410 Atomic Compare & Write Unit: 1 00:28:09.410 Fused Compare & Write: Not Supported 00:28:09.410 Scatter-Gather List 00:28:09.410 SGL Command Set: Supported 00:28:09.410 SGL Keyed: Not Supported 00:28:09.410 SGL Bit Bucket Descriptor: Not Supported 00:28:09.410 SGL Metadata Pointer: Not Supported 00:28:09.410 Oversized SGL: Not Supported 00:28:09.410 SGL Metadata Address: Not Supported 00:28:09.410 SGL Offset: Supported 00:28:09.410 Transport SGL Data Block: Not Supported 00:28:09.410 Replay Protected Memory Block: Not Supported 00:28:09.410 00:28:09.410 Firmware Slot Information 00:28:09.410 ========================= 00:28:09.410 Active slot: 0 00:28:09.410 00:28:09.410 Asymmetric Namespace Access 00:28:09.410 =========================== 00:28:09.410 Change Count : 0 00:28:09.410 Number of ANA Group Descriptors : 1 00:28:09.410 ANA Group Descriptor : 0 00:28:09.410 ANA Group ID : 1 00:28:09.410 Number of NSID Values : 1 00:28:09.410 Change Count : 0 00:28:09.410 ANA State : 1 00:28:09.410 Namespace Identifier : 1 00:28:09.410 00:28:09.410 Commands Supported and Effects 00:28:09.410 ============================== 00:28:09.410 Admin Commands 00:28:09.410 -------------- 00:28:09.410 Get Log Page (02h): Supported 00:28:09.410 Identify (06h): Supported 00:28:09.410 Abort (08h): Supported 00:28:09.410 Set Features (09h): Supported 00:28:09.410 Get Features (0Ah): Supported 00:28:09.410 Asynchronous Event Request (0Ch): Supported 00:28:09.410 Keep Alive (18h): Supported 00:28:09.410 I/O Commands 00:28:09.410 ------------ 00:28:09.410 Flush (00h): Supported 00:28:09.410 Write (01h): Supported LBA-Change 00:28:09.410 Read (02h): Supported 00:28:09.410 Write Zeroes (08h): Supported LBA-Change 00:28:09.410 Dataset Management (09h): Supported 00:28:09.410 00:28:09.410 Error Log 00:28:09.410 ========= 00:28:09.410 Entry: 0 00:28:09.410 Error Count: 0x3 00:28:09.410 Submission Queue Id: 0x0 00:28:09.410 Command Id: 0x5 00:28:09.410 Phase Bit: 0 00:28:09.410 Status Code: 0x2 00:28:09.410 Status Code Type: 0x0 00:28:09.410 Do Not Retry: 1 00:28:09.410 Error Location: 0x28 00:28:09.410 LBA: 0x0 00:28:09.410 Namespace: 0x0 00:28:09.410 Vendor Log Page: 0x0 00:28:09.410 ----------- 00:28:09.410 Entry: 1 00:28:09.410 Error Count: 0x2 00:28:09.410 Submission Queue Id: 0x0 00:28:09.410 Command Id: 0x5 00:28:09.410 Phase Bit: 0 00:28:09.410 Status Code: 0x2 00:28:09.410 Status Code Type: 0x0 00:28:09.410 Do Not Retry: 1 00:28:09.410 Error Location: 0x28 00:28:09.410 LBA: 0x0 00:28:09.410 Namespace: 0x0 00:28:09.410 Vendor Log Page: 0x0 00:28:09.410 ----------- 00:28:09.410 Entry: 2 00:28:09.410 Error Count: 0x1 00:28:09.410 Submission Queue Id: 0x0 00:28:09.410 Command Id: 0x4 00:28:09.410 Phase Bit: 0 00:28:09.411 Status Code: 0x2 00:28:09.411 Status Code Type: 0x0 00:28:09.411 Do Not Retry: 1 00:28:09.411 Error Location: 0x28 00:28:09.411 LBA: 0x0 00:28:09.411 Namespace: 0x0 00:28:09.411 Vendor Log Page: 0x0 00:28:09.411 00:28:09.411 Number of Queues 00:28:09.411 ================ 00:28:09.411 Number of I/O Submission Queues: 128 00:28:09.411 Number of I/O Completion Queues: 128 00:28:09.411 00:28:09.411 ZNS Specific Controller Data 00:28:09.411 ============================ 00:28:09.411 Zone Append Size Limit: 0 00:28:09.411 00:28:09.411 00:28:09.411 Active Namespaces 00:28:09.411 ================= 00:28:09.411 get_feature(0x05) failed 00:28:09.411 Namespace ID:1 00:28:09.411 Command Set Identifier: NVM (00h) 00:28:09.411 Deallocate: Supported 00:28:09.411 Deallocated/Unwritten Error: Not Supported 00:28:09.411 Deallocated Read Value: Unknown 00:28:09.411 Deallocate in Write Zeroes: Not Supported 00:28:09.411 Deallocated Guard Field: 0xFFFF 00:28:09.411 Flush: Supported 00:28:09.411 Reservation: Not Supported 00:28:09.411 Namespace Sharing Capabilities: Multiple Controllers 00:28:09.411 Size (in LBAs): 3750748848 (1788GiB) 00:28:09.411 Capacity (in LBAs): 3750748848 (1788GiB) 00:28:09.411 Utilization (in LBAs): 3750748848 (1788GiB) 00:28:09.411 UUID: fb0e7d3e-44ba-45ac-9d70-9f1b9272d46b 00:28:09.411 Thin Provisioning: Not Supported 00:28:09.411 Per-NS Atomic Units: Yes 00:28:09.411 Atomic Write Unit (Normal): 8 00:28:09.411 Atomic Write Unit (PFail): 8 00:28:09.411 Preferred Write Granularity: 8 00:28:09.411 Atomic Compare & Write Unit: 8 00:28:09.411 Atomic Boundary Size (Normal): 0 00:28:09.411 Atomic Boundary Size (PFail): 0 00:28:09.411 Atomic Boundary Offset: 0 00:28:09.411 NGUID/EUI64 Never Reused: No 00:28:09.411 ANA group ID: 1 00:28:09.411 Namespace Write Protected: No 00:28:09.411 Number of LBA Formats: 1 00:28:09.411 Current LBA Format: LBA Format #00 00:28:09.411 LBA Format #00: Data Size: 512 Metadata Size: 0 00:28:09.411 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # nvmftestfini 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@121 -- # sync 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@124 -- # set +e 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:28:09.411 rmmod nvme_tcp 00:28:09.411 rmmod nvme_fabrics 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@128 -- # set -e 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@129 -- # return 0 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@297 -- # iptr 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@791 -- # iptables-save 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@791 -- # iptables-restore 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:09.411 19:24:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # clean_kernel_target 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@712 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@714 -- # echo 0 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@716 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@718 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@719 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@721 -- # modules=(/sys/module/nvmet/holders/*) 00:28:11.956 19:24:23 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@723 -- # modprobe -r nvmet_tcp nvmet 00:28:11.956 19:24:24 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@726 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:28:16.165 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:28:16.165 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:28:16.165 00:28:16.165 real 0m21.344s 00:28:16.165 user 0m5.854s 00:28:16.165 sys 0m12.600s 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:28:16.165 ************************************ 00:28:16.165 END TEST nvmf_identify_kernel_target 00:28:16.165 ************************************ 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@30 -- # run_test nvmf_auth_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:28:16.165 ************************************ 00:28:16.165 START TEST nvmf_auth_host 00:28:16.165 ************************************ 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:28:16.165 * Looking for test storage... 00:28:16.165 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1693 -- # lcov --version 00:28:16.165 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # IFS=.-: 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # read -ra ver1 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # IFS=.-: 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # read -ra ver2 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@338 -- # local 'op=<' 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@340 -- # ver1_l=2 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@341 -- # ver2_l=1 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@344 -- # case "$op" in 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@345 -- # : 1 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # decimal 1 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=1 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 1 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # ver1[v]=1 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # decimal 2 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=2 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:28:16.166 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 2 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # ver2[v]=2 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # return 0 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:28:16.427 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:16.427 --rc genhtml_branch_coverage=1 00:28:16.427 --rc genhtml_function_coverage=1 00:28:16.427 --rc genhtml_legend=1 00:28:16.427 --rc geninfo_all_blocks=1 00:28:16.427 --rc geninfo_unexecuted_blocks=1 00:28:16.427 00:28:16.427 ' 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:28:16.427 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:16.427 --rc genhtml_branch_coverage=1 00:28:16.427 --rc genhtml_function_coverage=1 00:28:16.427 --rc genhtml_legend=1 00:28:16.427 --rc geninfo_all_blocks=1 00:28:16.427 --rc geninfo_unexecuted_blocks=1 00:28:16.427 00:28:16.427 ' 00:28:16.427 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:28:16.427 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:16.427 --rc genhtml_branch_coverage=1 00:28:16.427 --rc genhtml_function_coverage=1 00:28:16.427 --rc genhtml_legend=1 00:28:16.427 --rc geninfo_all_blocks=1 00:28:16.427 --rc geninfo_unexecuted_blocks=1 00:28:16.427 00:28:16.427 ' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:28:16.428 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:16.428 --rc genhtml_branch_coverage=1 00:28:16.428 --rc genhtml_function_coverage=1 00:28:16.428 --rc genhtml_legend=1 00:28:16.428 --rc geninfo_all_blocks=1 00:28:16.428 --rc geninfo_unexecuted_blocks=1 00:28:16.428 00:28:16.428 ' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # uname -s 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@15 -- # shopt -s extglob 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@5 -- # export PATH 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@51 -- # : 0 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:28:16.428 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@16 -- # dhgroups=("ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@17 -- # subnqn=nqn.2024-02.io.spdk:cnode0 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@18 -- # hostnqn=nqn.2024-02.io.spdk:host0 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@19 -- # nvmet_subsys=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@20 -- # nvmet_host=/sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # keys=() 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # ckeys=() 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@68 -- # nvmftestinit 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@476 -- # prepare_net_devs 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@438 -- # local -g is_hw=no 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # remove_spdk_ns 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@309 -- # xtrace_disable 00:28:16.428 19:24:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # pci_devs=() 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # net_devs=() 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # e810=() 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # local -ga e810 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # x722=() 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # local -ga x722 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # mlx=() 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # local -ga mlx 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:28:24.577 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:28:24.578 Found 0000:31:00.0 (0x8086 - 0x159b) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:28:24.578 Found 0000:31:00.1 (0x8086 - 0x159b) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:28:24.578 Found net devices under 0000:31:00.0: cvl_0_0 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:28:24.578 Found net devices under 0000:31:00.1: cvl_0_1 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # is_hw=yes 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:28:24.578 19:24:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:28:24.578 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:28:24.578 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:28:24.578 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.712 ms 00:28:24.578 00:28:24.578 --- 10.0.0.2 ping statistics --- 00:28:24.578 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:24.578 rtt min/avg/max/mdev = 0.712/0.712/0.712/0.000 ms 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:28:24.579 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:28:24.579 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.264 ms 00:28:24.579 00:28:24.579 --- 10.0.0.1 ping statistics --- 00:28:24.579 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:24.579 rtt min/avg/max/mdev = 0.264/0.264/0.264/0.000 ms 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@450 -- # return 0 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@69 -- # nvmfappstart -L nvme_auth 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@726 -- # xtrace_disable 00:28:24.579 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@509 -- # nvmfpid=4026803 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@510 -- # waitforlisten 4026803 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvme_auth 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # '[' -z 4026803 ']' 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:24.839 19:24:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@868 -- # return 0 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@732 -- # xtrace_disable 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@70 -- # trap 'cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log; cleanup' SIGINT SIGTERM EXIT 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key null 32 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=null 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=396a62129f09909ee78301c3b361b6b6 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.zO9 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 396a62129f09909ee78301c3b361b6b6 0 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 396a62129f09909ee78301c3b361b6b6 0 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=396a62129f09909ee78301c3b361b6b6 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=0 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.zO9 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.zO9 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # keys[0]=/tmp/spdk.key-null.zO9 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key sha512 64 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha512 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=64 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=d200b9fa224b0eba8f62112a617c334cb79bc63e2a8ecab02d571ae7e16abee6 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.NGX 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key d200b9fa224b0eba8f62112a617c334cb79bc63e2a8ecab02d571ae7e16abee6 3 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 d200b9fa224b0eba8f62112a617c334cb79bc63e2a8ecab02d571ae7e16abee6 3 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=d200b9fa224b0eba8f62112a617c334cb79bc63e2a8ecab02d571ae7e16abee6 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=3 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.NGX 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.NGX 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # ckeys[0]=/tmp/spdk.key-sha512.NGX 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key null 48 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=null 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=48 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=23b1b00882771e5fc0632d65e2c70676499afe64466e7a5a 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.mkd 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 23b1b00882771e5fc0632d65e2c70676499afe64466e7a5a 0 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 23b1b00882771e5fc0632d65e2c70676499afe64466e7a5a 0 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=23b1b00882771e5fc0632d65e2c70676499afe64466e7a5a 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=0 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.mkd 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.mkd 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # keys[1]=/tmp/spdk.key-null.mkd 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key sha384 48 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha384 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=48 00:28:25.783 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=9dc9e928c6513c780ec9a88b8eda2d24196f493fb87bc8f2 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.e29 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 9dc9e928c6513c780ec9a88b8eda2d24196f493fb87bc8f2 2 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 9dc9e928c6513c780ec9a88b8eda2d24196f493fb87bc8f2 2 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=9dc9e928c6513c780ec9a88b8eda2d24196f493fb87bc8f2 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=2 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.e29 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.e29 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # ckeys[1]=/tmp/spdk.key-sha384.e29 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha256 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=7fe918f98b6ae2a5fc6dbd782bc0d569 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.Wzv 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 7fe918f98b6ae2a5fc6dbd782bc0d569 1 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 7fe918f98b6ae2a5fc6dbd782bc0d569 1 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=7fe918f98b6ae2a5fc6dbd782bc0d569 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=1 00:28:25.784 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.Wzv 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.Wzv 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # keys[2]=/tmp/spdk.key-sha256.Wzv 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha256 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=95d156b9c3d97c9ce72fa72b67262581 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.kTS 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 95d156b9c3d97c9ce72fa72b67262581 1 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 95d156b9c3d97c9ce72fa72b67262581 1 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=95d156b9c3d97c9ce72fa72b67262581 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=1 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.kTS 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.kTS 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # ckeys[2]=/tmp/spdk.key-sha256.kTS 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key sha384 48 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha384 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=48 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=d2833e79e7a9ea279cad140da07b066e830586fee7f1eedb 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.866 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key d2833e79e7a9ea279cad140da07b066e830586fee7f1eedb 2 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 d2833e79e7a9ea279cad140da07b066e830586fee7f1eedb 2 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=d2833e79e7a9ea279cad140da07b066e830586fee7f1eedb 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=2 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.866 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.866 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # keys[3]=/tmp/spdk.key-sha384.866 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key null 32 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=null 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=1e3903d8c610469a8bfca2ca5543c6ef 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.YNv 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 1e3903d8c610469a8bfca2ca5543c6ef 0 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 1e3903d8c610469a8bfca2ca5543c6ef 0 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=1e3903d8c610469a8bfca2ca5543c6ef 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=0 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.YNv 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.YNv 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # ckeys[3]=/tmp/spdk.key-null.YNv 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # gen_dhchap_key sha512 64 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha512 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=64 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=0976c06c052be605a3a52a589a88a1e7353463453be8730af54aeda97a1b1b0a 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.wrC 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 0976c06c052be605a3a52a589a88a1e7353463453be8730af54aeda97a1b1b0a 3 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 0976c06c052be605a3a52a589a88a1e7353463453be8730af54aeda97a1b1b0a 3 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=0976c06c052be605a3a52a589a88a1e7353463453be8730af54aeda97a1b1b0a 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=3 00:28:26.046 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.wrC 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.wrC 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # keys[4]=/tmp/spdk.key-sha512.wrC 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # ckeys[4]= 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@79 -- # waitforlisten 4026803 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # '[' -z 4026803 ']' 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:26.307 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@868 -- # return 0 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.zO9 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha512.NGX ]] 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.NGX 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-null.mkd 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha384.e29 ]] 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.e29 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.307 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:28:26.308 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha256.Wzv 00:28:26.308 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.308 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.308 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.308 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha256.kTS ]] 00:28:26.308 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.kTS 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha384.866 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-null.YNv ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey3 /tmp/spdk.key-null.YNv 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key4 /tmp/spdk.key-sha512.wrC 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n '' ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@85 -- # nvmet_auth_init 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # get_main_ns_ip 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:26.569 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # configure_kernel_target nqn.2024-02.io.spdk:cnode0 10.0.0.1 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@660 -- # local kernel_name=nqn.2024-02.io.spdk:cnode0 kernel_target_ip=10.0.0.1 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@662 -- # nvmet=/sys/kernel/config/nvmet 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@663 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@664 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@665 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@667 -- # local block nvme 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@669 -- # [[ ! -e /sys/module/nvmet ]] 00:28:26.570 19:24:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@670 -- # modprobe nvmet 00:28:26.570 19:24:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet ]] 00:28:26.570 19:24:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@675 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:28:30.804 Waiting for block devices as requested 00:28:30.804 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:28:30.804 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:28:30.804 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:28:30.804 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:28:30.804 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:28:30.804 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:28:30.805 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:28:30.805 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:28:30.805 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:28:31.065 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:28:31.065 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:28:31.326 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:28:31.326 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:28:31.326 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:28:31.326 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:28:31.585 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:28:31.585 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@678 -- # for block in /sys/block/nvme* 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@679 -- # [[ -e /sys/block/nvme0n1 ]] 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@680 -- # is_block_zoned nvme0n1 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@681 -- # block_in_use nvme0n1 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:28:32.523 No valid GPT data, bailing 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # pt= 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@395 -- # return 1 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@681 -- # nvme=/dev/nvme0n1 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@684 -- # [[ -b /dev/nvme0n1 ]] 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:28:32.523 19:24:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@687 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@688 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@693 -- # echo SPDK-nqn.2024-02.io.spdk:cnode0 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@695 -- # echo 1 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@696 -- # echo /dev/nvme0n1 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@697 -- # echo 1 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@699 -- # echo 10.0.0.1 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@700 -- # echo tcp 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@701 -- # echo 4420 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@702 -- # echo ipv4 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@705 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 /sys/kernel/config/nvmet/ports/1/subsystems/ 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@708 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -a 10.0.0.1 -t tcp -s 4420 00:28:32.523 00:28:32.523 Discovery Log Number of Records 2, Generation counter 2 00:28:32.523 =====Discovery Log Entry 0====== 00:28:32.523 trtype: tcp 00:28:32.523 adrfam: ipv4 00:28:32.523 subtype: current discovery subsystem 00:28:32.523 treq: not specified, sq flow control disable supported 00:28:32.523 portid: 1 00:28:32.523 trsvcid: 4420 00:28:32.523 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:28:32.523 traddr: 10.0.0.1 00:28:32.523 eflags: none 00:28:32.523 sectype: none 00:28:32.523 =====Discovery Log Entry 1====== 00:28:32.523 trtype: tcp 00:28:32.523 adrfam: ipv4 00:28:32.523 subtype: nvme subsystem 00:28:32.523 treq: not specified, sq flow control disable supported 00:28:32.523 portid: 1 00:28:32.523 trsvcid: 4420 00:28:32.523 subnqn: nqn.2024-02.io.spdk:cnode0 00:28:32.523 traddr: 10.0.0.1 00:28:32.523 eflags: none 00:28:32.523 sectype: none 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@36 -- # mkdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@37 -- # echo 0 00:28:32.523 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@38 -- # ln -s /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@88 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s sha256,sha384,sha512 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # connect_authenticate sha256,sha384,sha512 ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 1 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256,sha384,sha512 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.524 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:32.784 nvme0n1 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 0 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 0 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.784 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.045 nvme0n1 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 1 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:33.045 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:33.046 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.046 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.307 nvme0n1 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 2 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.307 19:24:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.568 nvme0n1 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 3 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:33.568 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 3 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.569 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.830 nvme0n1 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 4 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 4 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.830 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.091 nvme0n1 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 0 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 0 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.091 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.352 nvme0n1 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 1 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:34.352 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 1 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.353 19:24:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.615 nvme0n1 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 2 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 2 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.615 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.877 nvme0n1 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 3 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 3 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.877 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.137 nvme0n1 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 4 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 4 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:35.137 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.138 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.399 nvme0n1 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 0 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 0 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.399 19:24:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.659 nvme0n1 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 1 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 1 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:35.659 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:35.918 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:35.918 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:35.918 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:35.918 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:35.918 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:35.918 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:35.918 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:35.919 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:35.919 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.919 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.179 nvme0n1 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 2 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 2 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.179 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.440 nvme0n1 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 3 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 3 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:36.440 19:24:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:36.440 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:36.440 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.440 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.701 nvme0n1 00:28:36.701 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.701 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:36.701 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.701 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:36.701 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.701 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 4 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 4 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.961 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.222 nvme0n1 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 0 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 0 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.222 19:24:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.793 nvme0n1 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 1 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 1 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.793 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.364 nvme0n1 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 2 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 2 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.364 19:24:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.936 nvme0n1 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.936 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 3 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 3 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.937 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:39.197 nvme0n1 00:28:39.197 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.197 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:39.197 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:39.197 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.197 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:39.197 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.457 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:39.457 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:39.457 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 4 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 4 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.458 19:24:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.029 nvme0n1 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 0 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 0 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.029 19:24:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.600 nvme0n1 00:28:40.600 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.600 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:40.600 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:40.600 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.600 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.600 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 1 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 1 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:40.860 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:40.861 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:40.861 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:40.861 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:40.861 19:24:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:41.429 nvme0n1 00:28:41.429 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.429 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:41.429 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.429 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:41.429 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:41.429 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.689 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 2 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 2 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.690 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:42.261 nvme0n1 00:28:42.261 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.261 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:42.261 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:42.261 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.261 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:42.261 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 3 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 3 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.522 19:24:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:43.094 nvme0n1 00:28:43.094 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.094 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:43.094 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:43.094 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.094 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:43.094 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 4 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 4 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.355 19:24:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:43.928 nvme0n1 00:28:43.928 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.928 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:43.928 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:43.929 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.929 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:43.929 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 0 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 0 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.191 nvme0n1 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.191 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 1 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 1 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.452 nvme0n1 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.452 19:24:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.452 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.452 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:44.452 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:44.452 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.452 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.452 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 2 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 2 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.453 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.714 nvme0n1 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 3 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 3 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.714 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.975 nvme0n1 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 4 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 4 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.975 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.236 nvme0n1 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 0 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 0 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.236 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.497 nvme0n1 00:28:45.497 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.497 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:45.497 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:45.497 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.497 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.497 19:24:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 1 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 1 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.497 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.758 nvme0n1 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 2 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 2 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:45.758 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.020 nvme0n1 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 3 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 3 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.020 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.281 nvme0n1 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 4 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:46.281 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 4 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.282 19:24:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.543 nvme0n1 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 0 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 0 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.543 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.803 nvme0n1 00:28:46.803 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.803 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:46.803 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:46.803 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.803 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.803 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 1 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 1 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:47.063 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.064 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.324 nvme0n1 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 2 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 2 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:47.324 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.325 19:24:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.586 nvme0n1 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 3 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 3 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.586 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:47.846 nvme0n1 00:28:47.846 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.846 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 4 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 4 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:48.107 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:48.108 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:48.108 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.108 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.368 nvme0n1 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:48.368 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 0 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 0 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.369 19:25:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.940 nvme0n1 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 1 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 1 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.940 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:49.512 nvme0n1 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 2 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:49.512 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 2 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:49.513 19:25:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:49.513 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.084 nvme0n1 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 3 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 3 00:28:50.084 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.085 19:25:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.657 nvme0n1 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 4 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:50.657 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 4 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.658 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:51.228 nvme0n1 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 0 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 0 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:28:51.228 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:51.229 19:25:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:51.800 nvme0n1 00:28:51.800 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:51.800 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:51.800 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:51.800 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:51.800 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:51.800 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 1 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 1 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.067 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.068 19:25:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:52.710 nvme0n1 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 2 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 2 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.710 19:25:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:53.685 nvme0n1 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 3 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 3 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:53.685 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:54.629 nvme0n1 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 4 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 4 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.629 19:25:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.203 nvme0n1 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 0 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 0 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:55.203 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.466 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.466 nvme0n1 00:28:55.466 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.466 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:55.466 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:55.466 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.466 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.466 19:25:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 1 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 1 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.466 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.727 nvme0n1 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 2 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 2 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:55.727 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.728 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.990 nvme0n1 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 3 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 3 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:55.990 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:55.991 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.991 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.252 nvme0n1 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 4 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 4 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.252 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.513 nvme0n1 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 0 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 0 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.513 19:25:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.513 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.775 nvme0n1 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 1 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 1 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.775 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.036 nvme0n1 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 2 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 2 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.036 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.298 nvme0n1 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 3 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 3 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.298 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.560 nvme0n1 00:28:57.560 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.560 19:25:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 4 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 4 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.560 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.821 nvme0n1 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 0 00:28:57.821 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 0 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.822 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.082 nvme0n1 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 1 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 1 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.082 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.343 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.605 nvme0n1 00:28:58.605 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.605 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:58.605 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:58.605 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.605 19:25:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 2 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 2 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.605 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.866 nvme0n1 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 3 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 3 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.866 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.127 nvme0n1 00:28:59.127 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.127 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:59.127 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:59.127 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.127 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.127 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.388 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:59.388 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:59.388 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.388 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.388 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.388 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 4 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 4 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.389 19:25:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.649 nvme0n1 00:28:59.649 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.649 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:28:59.649 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 0 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 0 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.650 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.223 nvme0n1 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 1 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 1 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:00.223 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:00.224 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:00.224 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:00.224 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:29:00.224 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.224 19:25:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.796 nvme0n1 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:00.796 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 2 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 2 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.797 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.370 nvme0n1 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 3 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 3 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.370 19:25:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.942 nvme0n1 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 4 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 4 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.942 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:02.202 nvme0n1 00:29:02.202 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:02.202 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:02.202 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:02.202 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:02.202 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 0 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:Mzk2YTYyMTI5ZjA5OTA5ZWU3ODMwMWMzYjM2MWI2YjawxKXf: 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZDIwMGI5ZmEyMjRiMGViYThmNjIxMTJhNjE3YzMzNGNiNzliYzYzZTJhOGVjYWIwMmQ1NzFhZTdlMTZhYmVlNvNpfYo=: 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 0 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:02.463 19:25:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:03.035 nvme0n1 00:29:03.035 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.035 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:03.035 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:03.035 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.035 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 1 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 1 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.297 19:25:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:03.869 nvme0n1 00:29:03.869 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.869 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:03.869 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.869 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:03.869 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 2 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 2 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.131 19:25:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:04.704 nvme0n1 00:29:04.704 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.704 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:04.704 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:04.704 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.704 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 3 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDI4MzNlNzllN2E5ZWEyNzljYWQxNDBkYTA3YjA2NmU4MzA1ODZmZWU3ZjFlZWRi/5PcRg==: 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: ]] 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUzOTAzZDhjNjEwNDY5YThiZmNhMmNhNTU0M2M2ZWaq/yJi: 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 3 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:04.965 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:04.966 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:04.966 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:04.966 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:04.966 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:04.966 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:29:04.966 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.966 19:25:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:05.537 nvme0n1 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 4 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MDk3NmMwNmMwNTJiZTYwNWEzYTUyYTU4OWE4OGExZTczNTM0NjM0NTNiZTg3MzBhZjU0YWVkYTk3YTFiMWIwYXYVeIU=: 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 4 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.798 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.370 nvme0n1 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:06.370 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:29:06.631 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:29:06.631 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:06.631 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:29:06.631 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:06.631 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@111 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:29:06.631 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.631 19:25:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # get_main_ns_ip 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.631 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.631 request: 00:29:06.631 { 00:29:06.631 "name": "nvme0", 00:29:06.631 "trtype": "tcp", 00:29:06.631 "traddr": "10.0.0.1", 00:29:06.631 "adrfam": "ipv4", 00:29:06.631 "trsvcid": "4420", 00:29:06.631 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:29:06.631 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:29:06.631 "prchk_reftag": false, 00:29:06.631 "prchk_guard": false, 00:29:06.631 "hdgst": false, 00:29:06.631 "ddgst": false, 00:29:06.631 "allow_unrecognized_csi": false, 00:29:06.631 "method": "bdev_nvme_attach_controller", 00:29:06.631 "req_id": 1 00:29:06.631 } 00:29:06.631 Got JSON-RPC error response 00:29:06.631 response: 00:29:06.631 { 00:29:06.632 "code": -5, 00:29:06.632 "message": "Input/output error" 00:29:06.632 } 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # rpc_cmd bdev_nvme_get_controllers 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # jq length 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # (( 0 == 0 )) 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # get_main_ns_ip 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.632 request: 00:29:06.632 { 00:29:06.632 "name": "nvme0", 00:29:06.632 "trtype": "tcp", 00:29:06.632 "traddr": "10.0.0.1", 00:29:06.632 "adrfam": "ipv4", 00:29:06.632 "trsvcid": "4420", 00:29:06.632 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:29:06.632 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:29:06.632 "prchk_reftag": false, 00:29:06.632 "prchk_guard": false, 00:29:06.632 "hdgst": false, 00:29:06.632 "ddgst": false, 00:29:06.632 "dhchap_key": "key2", 00:29:06.632 "allow_unrecognized_csi": false, 00:29:06.632 "method": "bdev_nvme_attach_controller", 00:29:06.632 "req_id": 1 00:29:06.632 } 00:29:06.632 Got JSON-RPC error response 00:29:06.632 response: 00:29:06.632 { 00:29:06.632 "code": -5, 00:29:06.632 "message": "Input/output error" 00:29:06.632 } 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # rpc_cmd bdev_nvme_get_controllers 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # jq length 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # (( 0 == 0 )) 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # get_main_ns_ip 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:06.632 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.893 request: 00:29:06.893 { 00:29:06.893 "name": "nvme0", 00:29:06.893 "trtype": "tcp", 00:29:06.893 "traddr": "10.0.0.1", 00:29:06.893 "adrfam": "ipv4", 00:29:06.893 "trsvcid": "4420", 00:29:06.893 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:29:06.893 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:29:06.893 "prchk_reftag": false, 00:29:06.893 "prchk_guard": false, 00:29:06.893 "hdgst": false, 00:29:06.893 "ddgst": false, 00:29:06.893 "dhchap_key": "key1", 00:29:06.893 "dhchap_ctrlr_key": "ckey2", 00:29:06.893 "allow_unrecognized_csi": false, 00:29:06.893 "method": "bdev_nvme_attach_controller", 00:29:06.893 "req_id": 1 00:29:06.893 } 00:29:06.893 Got JSON-RPC error response 00:29:06.893 response: 00:29:06.893 { 00:29:06.893 "code": -5, 00:29:06.893 "message": "Input/output error" 00:29:06.893 } 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # get_main_ns_ip 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:06.893 nvme0n1 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@132 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@133 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.893 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # rpc_cmd bdev_nvme_get_controllers 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # jq -r '.[].name' 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@136 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:07.155 request: 00:29:07.155 { 00:29:07.155 "name": "nvme0", 00:29:07.155 "dhchap_key": "key1", 00:29:07.155 "dhchap_ctrlr_key": "ckey2", 00:29:07.155 "method": "bdev_nvme_set_keys", 00:29:07.155 "req_id": 1 00:29:07.155 } 00:29:07.155 Got JSON-RPC error response 00:29:07.155 response: 00:29:07.155 { 00:29:07.155 "code": -13, 00:29:07.155 "message": "Permission denied" 00:29:07.155 } 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 1 != 0 )) 00:29:07.155 19:25:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@138 -- # sleep 1s 00:29:08.541 19:25:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:29:08.541 19:25:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:29:08.541 19:25:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.541 19:25:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:08.541 19:25:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.541 19:25:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 1 != 0 )) 00:29:08.541 19:25:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@138 -- # sleep 1s 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 0 != 0 )) 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@141 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MjNiMWIwMDg4Mjc3MWU1ZmMwNjMyZDY1ZTJjNzA2NzY0OTlhZmU2NDQ2NmU3YTVhctiYZg==: 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: ]] 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OWRjOWU5MjhjNjUxM2M3ODBlYzlhODhiOGVkYTJkMjQxOTZmNDkzZmI4N2JjOGYyT88tNA==: 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # get_main_ns_ip 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:29:09.489 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:09.490 nvme0n1 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@146 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:N2ZlOTE4Zjk4YjZhZTJhNWZjNmRiZDc4MmJjMGQ1Njn4QNOl: 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: ]] 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:OTVkMTU2YjljM2Q5N2M5Y2U3MmZhNzJiNjcyNjI1ODHEJdZe: 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@147 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:29:09.490 19:25:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:09.490 request: 00:29:09.490 { 00:29:09.490 "name": "nvme0", 00:29:09.490 "dhchap_key": "key2", 00:29:09.490 "dhchap_ctrlr_key": "ckey1", 00:29:09.490 "method": "bdev_nvme_set_keys", 00:29:09.490 "req_id": 1 00:29:09.490 } 00:29:09.490 Got JSON-RPC error response 00:29:09.490 response: 00:29:09.490 { 00:29:09.490 "code": -13, 00:29:09.490 "message": "Permission denied" 00:29:09.490 } 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:09.490 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.751 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 1 != 0 )) 00:29:09.751 19:25:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@149 -- # sleep 1s 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 0 != 0 )) 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@152 -- # trap - SIGINT SIGTERM EXIT 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@153 -- # cleanup 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@24 -- # nvmftestfini 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@516 -- # nvmfcleanup 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@121 -- # sync 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@124 -- # set +e 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:29:10.694 rmmod nvme_tcp 00:29:10.694 rmmod nvme_fabrics 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@128 -- # set -e 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@129 -- # return 0 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@517 -- # '[' -n 4026803 ']' 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@518 -- # killprocess 4026803 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@954 -- # '[' -z 4026803 ']' 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@958 -- # kill -0 4026803 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@959 -- # uname 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4026803 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4026803' 00:29:10.694 killing process with pid 4026803 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@973 -- # kill 4026803 00:29:10.694 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@978 -- # wait 4026803 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@297 -- # iptr 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@791 -- # iptables-save 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@791 -- # iptables-restore 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:10.956 19:25:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@25 -- # rm /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@26 -- # rmdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@27 -- # clean_kernel_target 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@712 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 ]] 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@714 -- # echo 0 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@716 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2024-02.io.spdk:cnode0 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@718 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@719 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@721 -- # modules=(/sys/module/nvmet/holders/*) 00:29:12.874 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@723 -- # modprobe -r nvmet_tcp nvmet 00:29:13.134 19:25:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@726 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:29:17.343 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:29:17.343 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:29:17.605 19:25:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@28 -- # rm -f /tmp/spdk.key-null.zO9 /tmp/spdk.key-null.mkd /tmp/spdk.key-sha256.Wzv /tmp/spdk.key-sha384.866 /tmp/spdk.key-sha512.wrC /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log 00:29:17.605 19:25:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:29:21.814 0000:80:01.6 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:80:01.7 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:80:01.4 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:80:01.5 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:80:01.2 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:80:01.3 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:80:01.0 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:80:01.1 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:00:01.6 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:65:00.0 (144d a80a): Already using the vfio-pci driver 00:29:21.814 0000:00:01.7 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:00:01.4 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:00:01.5 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:00:01.2 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:00:01.3 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:00:01.0 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 0000:00:01.1 (8086 0b00): Already using the vfio-pci driver 00:29:21.814 00:29:21.814 real 1m5.612s 00:29:21.814 user 0m58.156s 00:29:21.814 sys 0m17.499s 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:29:21.814 ************************************ 00:29:21.814 END TEST nvmf_auth_host 00:29:21.814 ************************************ 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@32 -- # [[ tcp == \t\c\p ]] 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@33 -- # run_test nvmf_digest /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:29:21.814 ************************************ 00:29:21.814 START TEST nvmf_digest 00:29:21.814 ************************************ 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:29:21.814 * Looking for test storage... 00:29:21.814 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1693 -- # lcov --version 00:29:21.814 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@333 -- # local ver1 ver1_l 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@334 -- # local ver2 ver2_l 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # IFS=.-: 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # read -ra ver1 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # IFS=.-: 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # read -ra ver2 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@338 -- # local 'op=<' 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@340 -- # ver1_l=2 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@341 -- # ver2_l=1 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:29:22.075 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@344 -- # case "$op" in 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@345 -- # : 1 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v = 0 )) 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # decimal 1 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=1 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 1 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # ver1[v]=1 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # decimal 2 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=2 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 2 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # ver2[v]=2 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # return 0 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:29:22.076 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.076 --rc genhtml_branch_coverage=1 00:29:22.076 --rc genhtml_function_coverage=1 00:29:22.076 --rc genhtml_legend=1 00:29:22.076 --rc geninfo_all_blocks=1 00:29:22.076 --rc geninfo_unexecuted_blocks=1 00:29:22.076 00:29:22.076 ' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:29:22.076 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.076 --rc genhtml_branch_coverage=1 00:29:22.076 --rc genhtml_function_coverage=1 00:29:22.076 --rc genhtml_legend=1 00:29:22.076 --rc geninfo_all_blocks=1 00:29:22.076 --rc geninfo_unexecuted_blocks=1 00:29:22.076 00:29:22.076 ' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:29:22.076 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.076 --rc genhtml_branch_coverage=1 00:29:22.076 --rc genhtml_function_coverage=1 00:29:22.076 --rc genhtml_legend=1 00:29:22.076 --rc geninfo_all_blocks=1 00:29:22.076 --rc geninfo_unexecuted_blocks=1 00:29:22.076 00:29:22.076 ' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:29:22.076 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.076 --rc genhtml_branch_coverage=1 00:29:22.076 --rc genhtml_function_coverage=1 00:29:22.076 --rc genhtml_legend=1 00:29:22.076 --rc geninfo_all_blocks=1 00:29:22.076 --rc geninfo_unexecuted_blocks=1 00:29:22.076 00:29:22.076 ' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # uname -s 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@15 -- # shopt -s extglob 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@5 -- # export PATH 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@51 -- # : 0 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:29:22.076 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@55 -- # have_pci_nics=0 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@14 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@15 -- # bperfsock=/var/tmp/bperf.sock 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@16 -- # runtime=2 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@136 -- # [[ tcp != \t\c\p ]] 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@138 -- # nvmftestinit 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@476 -- # prepare_net_devs 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@438 -- # local -g is_hw=no 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # remove_spdk_ns 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@309 -- # xtrace_disable 00:29:22.076 19:25:34 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # pci_devs=() 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # local -a pci_devs 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # pci_net_devs=() 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # pci_drivers=() 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # local -A pci_drivers 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # net_devs=() 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # local -ga net_devs 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # e810=() 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # local -ga e810 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # x722=() 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # local -ga x722 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # mlx=() 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # local -ga mlx 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:29:30.217 Found 0000:31:00.0 (0x8086 - 0x159b) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:29:30.217 Found 0000:31:00.1 (0x8086 - 0x159b) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@418 -- # [[ up == up ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:29:30.217 Found net devices under 0000:31:00.0: cvl_0_0 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@418 -- # [[ up == up ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:29:30.217 Found net devices under 0000:31:00.1: cvl_0_1 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # is_hw=yes 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:29:30.217 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:29:30.218 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:29:30.218 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:29:30.218 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:29:30.218 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:29:30.479 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:29:30.479 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.509 ms 00:29:30.479 00:29:30.479 --- 10.0.0.2 ping statistics --- 00:29:30.479 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:30.479 rtt min/avg/max/mdev = 0.509/0.509/0.509/0.000 ms 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:29:30.479 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:29:30.479 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.302 ms 00:29:30.479 00:29:30.479 --- 10.0.0.1 ping statistics --- 00:29:30.479 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:30.479 rtt min/avg/max/mdev = 0.302/0.302/0.302/0.000 ms 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@450 -- # return 0 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@140 -- # trap cleanup SIGINT SIGTERM EXIT 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@141 -- # [[ 0 -eq 1 ]] 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@145 -- # run_test nvmf_digest_clean run_digest 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:29:30.479 ************************************ 00:29:30.479 START TEST nvmf_digest_clean 00:29:30.479 ************************************ 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1129 -- # run_digest 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@120 -- # local dsa_initiator 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # [[ '' == \d\s\a\_\i\n\i\t\i\a\t\o\r ]] 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # dsa_initiator=false 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@123 -- # tgt_params=("--wait-for-rpc") 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@124 -- # nvmfappstart --wait-for-rpc 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@726 -- # xtrace_disable 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@509 -- # nvmfpid=4045612 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@510 -- # waitforlisten 4045612 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 4045612 ']' 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:30.479 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:30.480 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:30.480 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:30.480 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:30.480 19:25:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:30.480 [2024-11-26 19:25:42.998166] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:30.480 [2024-11-26 19:25:42.998226] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:30.480 [2024-11-26 19:25:43.088533] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:30.740 [2024-11-26 19:25:43.128846] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:30.740 [2024-11-26 19:25:43.128888] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:30.740 [2024-11-26 19:25:43.128896] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:30.740 [2024-11-26 19:25:43.128903] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:30.740 [2024-11-26 19:25:43.128909] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:30.740 [2024-11-26 19:25:43.129543] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@732 -- # xtrace_disable 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@125 -- # [[ '' == \d\s\a\_\t\a\r\g\e\t ]] 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@126 -- # common_target_config 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@43 -- # rpc_cmd 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.310 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:31.310 null0 00:29:31.310 [2024-11-26 19:25:43.907079] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:31.310 [2024-11-26 19:25:43.931287] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@128 -- # run_bperf randread 4096 128 false 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=4045778 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 4045778 /var/tmp/bperf.sock 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 4045778 ']' 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:31.570 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:31.570 19:25:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:31.570 [2024-11-26 19:25:43.988276] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:31.570 [2024-11-26 19:25:43.988325] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4045778 ] 00:29:31.570 [2024-11-26 19:25:44.084162] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:31.570 [2024-11-26 19:25:44.120103] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:32.513 19:25:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:32.513 19:25:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:29:32.513 19:25:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:29:32.513 19:25:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:29:32.513 19:25:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:29:32.513 19:25:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:32.513 19:25:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:32.773 nvme0n1 00:29:32.773 19:25:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:29:32.773 19:25:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:32.773 Running I/O for 2 seconds... 00:29:35.096 18168.00 IOPS, 70.97 MiB/s [2024-11-26T18:25:47.721Z] 18588.50 IOPS, 72.61 MiB/s 00:29:35.096 Latency(us) 00:29:35.096 [2024-11-26T18:25:47.721Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:35.096 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:29:35.096 nvme0n1 : 2.01 18618.30 72.73 0.00 0.00 6865.27 3126.61 18568.53 00:29:35.096 [2024-11-26T18:25:47.721Z] =================================================================================================================== 00:29:35.096 [2024-11-26T18:25:47.721Z] Total : 18618.30 72.73 0.00 0.00 6865.27 3126.61 18568.53 00:29:35.096 { 00:29:35.096 "results": [ 00:29:35.096 { 00:29:35.096 "job": "nvme0n1", 00:29:35.096 "core_mask": "0x2", 00:29:35.096 "workload": "randread", 00:29:35.096 "status": "finished", 00:29:35.096 "queue_depth": 128, 00:29:35.096 "io_size": 4096, 00:29:35.096 "runtime": 2.007058, 00:29:35.096 "iops": 18618.29603329849, 00:29:35.096 "mibps": 72.72771888007223, 00:29:35.096 "io_failed": 0, 00:29:35.096 "io_timeout": 0, 00:29:35.096 "avg_latency_us": 6865.270180546635, 00:29:35.096 "min_latency_us": 3126.6133333333332, 00:29:35.096 "max_latency_us": 18568.533333333333 00:29:35.096 } 00:29:35.096 ], 00:29:35.096 "core_count": 1 00:29:35.096 } 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:29:35.096 | select(.opcode=="crc32c") 00:29:35.096 | "\(.module_name) \(.executed)"' 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 4045778 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 4045778 ']' 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 4045778 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4045778 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:29:35.096 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:29:35.097 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4045778' 00:29:35.097 killing process with pid 4045778 00:29:35.097 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 4045778 00:29:35.097 Received shutdown signal, test time was about 2.000000 seconds 00:29:35.097 00:29:35.097 Latency(us) 00:29:35.097 [2024-11-26T18:25:47.722Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:35.097 [2024-11-26T18:25:47.722Z] =================================================================================================================== 00:29:35.097 [2024-11-26T18:25:47.722Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:35.097 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 4045778 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@129 -- # run_bperf randread 131072 16 false 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=4046460 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 4046460 /var/tmp/bperf.sock 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 4046460 ']' 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:35.357 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:35.357 19:25:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:35.357 [2024-11-26 19:25:47.774504] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:35.357 [2024-11-26 19:25:47.774559] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4046460 ] 00:29:35.357 I/O size of 131072 is greater than zero copy threshold (65536). 00:29:35.357 Zero copy mechanism will not be used. 00:29:35.357 [2024-11-26 19:25:47.854968] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:35.357 [2024-11-26 19:25:47.884692] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:36.299 19:25:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:36.300 19:25:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:29:36.300 19:25:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:29:36.300 19:25:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:29:36.300 19:25:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:29:36.300 19:25:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:36.300 19:25:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:36.560 nvme0n1 00:29:36.560 19:25:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:29:36.560 19:25:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:36.560 I/O size of 131072 is greater than zero copy threshold (65536). 00:29:36.560 Zero copy mechanism will not be used. 00:29:36.560 Running I/O for 2 seconds... 00:29:38.885 2952.00 IOPS, 369.00 MiB/s [2024-11-26T18:25:51.510Z] 2966.00 IOPS, 370.75 MiB/s 00:29:38.885 Latency(us) 00:29:38.885 [2024-11-26T18:25:51.510Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:38.885 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:29:38.885 nvme0n1 : 2.00 2967.58 370.95 0.00 0.00 5389.31 1017.17 12834.13 00:29:38.885 [2024-11-26T18:25:51.510Z] =================================================================================================================== 00:29:38.885 [2024-11-26T18:25:51.510Z] Total : 2967.58 370.95 0.00 0.00 5389.31 1017.17 12834.13 00:29:38.885 { 00:29:38.885 "results": [ 00:29:38.885 { 00:29:38.885 "job": "nvme0n1", 00:29:38.885 "core_mask": "0x2", 00:29:38.885 "workload": "randread", 00:29:38.885 "status": "finished", 00:29:38.885 "queue_depth": 16, 00:29:38.885 "io_size": 131072, 00:29:38.885 "runtime": 2.004328, 00:29:38.885 "iops": 2967.578160859899, 00:29:38.885 "mibps": 370.9472701074874, 00:29:38.885 "io_failed": 0, 00:29:38.885 "io_timeout": 0, 00:29:38.885 "avg_latency_us": 5389.308442053351, 00:29:38.885 "min_latency_us": 1017.1733333333333, 00:29:38.885 "max_latency_us": 12834.133333333333 00:29:38.885 } 00:29:38.885 ], 00:29:38.885 "core_count": 1 00:29:38.885 } 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:29:38.885 | select(.opcode=="crc32c") 00:29:38.885 | "\(.module_name) \(.executed)"' 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 4046460 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 4046460 ']' 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 4046460 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4046460 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4046460' 00:29:38.885 killing process with pid 4046460 00:29:38.885 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 4046460 00:29:38.885 Received shutdown signal, test time was about 2.000000 seconds 00:29:38.885 00:29:38.885 Latency(us) 00:29:38.885 [2024-11-26T18:25:51.510Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:38.885 [2024-11-26T18:25:51.511Z] =================================================================================================================== 00:29:38.886 [2024-11-26T18:25:51.511Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 4046460 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@130 -- # run_bperf randwrite 4096 128 false 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=4047183 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 4047183 /var/tmp/bperf.sock 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 4047183 ']' 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:38.886 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:38.886 19:25:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:39.146 [2024-11-26 19:25:51.524691] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:39.146 [2024-11-26 19:25:51.524750] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4047183 ] 00:29:39.146 [2024-11-26 19:25:51.613463] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:39.146 [2024-11-26 19:25:51.643169] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:39.719 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:39.719 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:29:39.719 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:29:39.719 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:29:39.719 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:29:39.980 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:39.980 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:40.551 nvme0n1 00:29:40.551 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:29:40.552 19:25:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:40.552 Running I/O for 2 seconds... 00:29:42.478 21446.00 IOPS, 83.77 MiB/s [2024-11-26T18:25:55.103Z] 21566.50 IOPS, 84.24 MiB/s 00:29:42.478 Latency(us) 00:29:42.478 [2024-11-26T18:25:55.103Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:42.478 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:29:42.478 nvme0n1 : 2.01 21572.08 84.27 0.00 0.00 5925.61 3577.17 13544.11 00:29:42.478 [2024-11-26T18:25:55.103Z] =================================================================================================================== 00:29:42.478 [2024-11-26T18:25:55.103Z] Total : 21572.08 84.27 0.00 0.00 5925.61 3577.17 13544.11 00:29:42.478 { 00:29:42.478 "results": [ 00:29:42.478 { 00:29:42.478 "job": "nvme0n1", 00:29:42.478 "core_mask": "0x2", 00:29:42.478 "workload": "randwrite", 00:29:42.478 "status": "finished", 00:29:42.478 "queue_depth": 128, 00:29:42.478 "io_size": 4096, 00:29:42.478 "runtime": 2.005416, 00:29:42.478 "iops": 21572.0827997782, 00:29:42.478 "mibps": 84.2659484366336, 00:29:42.478 "io_failed": 0, 00:29:42.478 "io_timeout": 0, 00:29:42.478 "avg_latency_us": 5925.612627848023, 00:29:42.478 "min_latency_us": 3577.173333333333, 00:29:42.478 "max_latency_us": 13544.106666666667 00:29:42.478 } 00:29:42.478 ], 00:29:42.478 "core_count": 1 00:29:42.478 } 00:29:42.478 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:29:42.478 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:29:42.478 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:29:42.478 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:29:42.478 | select(.opcode=="crc32c") 00:29:42.478 | "\(.module_name) \(.executed)"' 00:29:42.478 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 4047183 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 4047183 ']' 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 4047183 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4047183 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4047183' 00:29:42.780 killing process with pid 4047183 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 4047183 00:29:42.780 Received shutdown signal, test time was about 2.000000 seconds 00:29:42.780 00:29:42.780 Latency(us) 00:29:42.780 [2024-11-26T18:25:55.405Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:42.780 [2024-11-26T18:25:55.405Z] =================================================================================================================== 00:29:42.780 [2024-11-26T18:25:55.405Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:42.780 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 4047183 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@131 -- # run_bperf randwrite 131072 16 false 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=4048038 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 4048038 /var/tmp/bperf.sock 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 4048038 ']' 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:43.045 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:43.045 19:25:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:43.045 [2024-11-26 19:25:55.441153] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:43.045 [2024-11-26 19:25:55.441212] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4048038 ] 00:29:43.045 I/O size of 131072 is greater than zero copy threshold (65536). 00:29:43.045 Zero copy mechanism will not be used. 00:29:43.045 [2024-11-26 19:25:55.529306] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:43.045 [2024-11-26 19:25:55.558736] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:43.615 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:43.616 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:29:43.616 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:29:43.616 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:29:43.616 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:29:43.876 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:43.876 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:44.137 nvme0n1 00:29:44.137 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:29:44.137 19:25:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:44.397 I/O size of 131072 is greater than zero copy threshold (65536). 00:29:44.397 Zero copy mechanism will not be used. 00:29:44.397 Running I/O for 2 seconds... 00:29:46.279 4480.00 IOPS, 560.00 MiB/s [2024-11-26T18:25:58.904Z] 4316.50 IOPS, 539.56 MiB/s 00:29:46.279 Latency(us) 00:29:46.279 [2024-11-26T18:25:58.904Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:46.279 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:29:46.279 nvme0n1 : 2.01 4313.46 539.18 0.00 0.00 3702.80 1583.79 7973.55 00:29:46.279 [2024-11-26T18:25:58.904Z] =================================================================================================================== 00:29:46.279 [2024-11-26T18:25:58.904Z] Total : 4313.46 539.18 0.00 0.00 3702.80 1583.79 7973.55 00:29:46.279 { 00:29:46.279 "results": [ 00:29:46.279 { 00:29:46.279 "job": "nvme0n1", 00:29:46.279 "core_mask": "0x2", 00:29:46.279 "workload": "randwrite", 00:29:46.279 "status": "finished", 00:29:46.279 "queue_depth": 16, 00:29:46.279 "io_size": 131072, 00:29:46.279 "runtime": 2.005815, 00:29:46.279 "iops": 4313.458619065068, 00:29:46.279 "mibps": 539.1823273831335, 00:29:46.279 "io_failed": 0, 00:29:46.279 "io_timeout": 0, 00:29:46.279 "avg_latency_us": 3702.7971767606723, 00:29:46.279 "min_latency_us": 1583.7866666666666, 00:29:46.279 "max_latency_us": 7973.546666666667 00:29:46.279 } 00:29:46.279 ], 00:29:46.279 "core_count": 1 00:29:46.279 } 00:29:46.279 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:29:46.279 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:29:46.279 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:29:46.279 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:29:46.279 | select(.opcode=="crc32c") 00:29:46.279 | "\(.module_name) \(.executed)"' 00:29:46.279 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 4048038 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 4048038 ']' 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 4048038 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:46.540 19:25:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4048038 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4048038' 00:29:46.540 killing process with pid 4048038 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 4048038 00:29:46.540 Received shutdown signal, test time was about 2.000000 seconds 00:29:46.540 00:29:46.540 Latency(us) 00:29:46.540 [2024-11-26T18:25:59.165Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:46.540 [2024-11-26T18:25:59.165Z] =================================================================================================================== 00:29:46.540 [2024-11-26T18:25:59.165Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 4048038 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@132 -- # killprocess 4045612 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 4045612 ']' 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 4045612 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:46.540 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4045612 00:29:46.799 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:46.799 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:46.799 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4045612' 00:29:46.799 killing process with pid 4045612 00:29:46.799 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 4045612 00:29:46.799 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 4045612 00:29:46.799 00:29:46.799 real 0m16.401s 00:29:46.799 user 0m32.506s 00:29:46.800 sys 0m3.534s 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:29:46.800 ************************************ 00:29:46.800 END TEST nvmf_digest_clean 00:29:46.800 ************************************ 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@147 -- # run_test nvmf_digest_error run_digest_error 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:29:46.800 ************************************ 00:29:46.800 START TEST nvmf_digest_error 00:29:46.800 ************************************ 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1129 -- # run_digest_error 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@102 -- # nvmfappstart --wait-for-rpc 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@726 -- # xtrace_disable 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@509 -- # nvmfpid=4048865 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@510 -- # waitforlisten 4048865 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 4048865 ']' 00:29:46.800 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:47.059 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:47.060 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:47.060 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:47.060 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:47.060 19:25:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:47.060 [2024-11-26 19:25:59.485284] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:47.060 [2024-11-26 19:25:59.485339] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:47.060 [2024-11-26 19:25:59.574199] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:47.060 [2024-11-26 19:25:59.614443] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:47.060 [2024-11-26 19:25:59.614481] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:47.060 [2024-11-26 19:25:59.614489] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:47.060 [2024-11-26 19:25:59.614496] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:47.060 [2024-11-26 19:25:59.614502] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:47.060 [2024-11-26 19:25:59.615108] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@732 -- # xtrace_disable 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@104 -- # rpc_cmd accel_assign_opc -o crc32c -m error 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:48.004 [2024-11-26 19:26:00.329159] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation crc32c will be assigned to module error 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@105 -- # common_target_config 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@43 -- # rpc_cmd 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:48.004 null0 00:29:48.004 [2024-11-26 19:26:00.412451] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:48.004 [2024-11-26 19:26:00.436680] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@108 -- # run_bperf_err randread 4096 128 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=4048979 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 4048979 /var/tmp/bperf.sock 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 4048979 ']' 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:48.004 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:48.004 19:26:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:48.004 [2024-11-26 19:26:00.493906] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:48.004 [2024-11-26 19:26:00.493955] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4048979 ] 00:29:48.004 [2024-11-26 19:26:00.581755] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:48.004 [2024-11-26 19:26:00.612104] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:48.946 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:49.207 nvme0n1 00:29:49.207 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:29:49.207 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:49.207 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:49.207 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:49.207 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:29:49.207 19:26:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:49.207 Running I/O for 2 seconds... 00:29:49.207 [2024-11-26 19:26:01.813020] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.207 [2024-11-26 19:26:01.813054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:4973 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.207 [2024-11-26 19:26:01.813063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.207 [2024-11-26 19:26:01.827640] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.207 [2024-11-26 19:26:01.827660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:19993 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.207 [2024-11-26 19:26:01.827667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.841464] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.841487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:10884 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.841494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.854896] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.854914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:13933 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.854921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.867976] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.867993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:12015 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.868000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.880949] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.880968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:25331 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.880975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.893212] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.893234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:24867 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.893241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.906371] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.906389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:4143 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.906395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.918501] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.918519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:10074 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.918525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.930132] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.930150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:12314 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.930156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.941537] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.941554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:9400 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.941561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.954529] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.954547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:4816 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.954554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.968925] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.968943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:15234 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.968950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.979834] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.979852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:15945 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.979858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:01.993892] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:01.993910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:13900 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:01.993917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:02.005488] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:02.005510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:15619 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:02.005517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:02.020193] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:02.020211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:10279 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:02.020219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:02.033142] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:02.033160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:23473 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:02.033167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:02.046471] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:02.046489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:21464 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:02.046496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:02.059757] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:02.059775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:8686 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:02.059782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:02.072315] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:02.072332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:13819 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:02.072339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:52 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.469 [2024-11-26 19:26:02.086806] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.469 [2024-11-26 19:26:02.086824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:4589 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.469 [2024-11-26 19:26:02.086831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.101362] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.101381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:21060 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.101387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.114675] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.114693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:17191 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.114703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.127156] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.127173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:6048 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.127180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.139688] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.139705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:4107 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.139712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.151836] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.151854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:15212 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.151861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.163098] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.163116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:10930 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.163123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.176744] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.176763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:935 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.176769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.189294] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.189312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:7305 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.189319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.201401] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.201420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4121 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.201427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.214761] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.214781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:6406 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.214788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.228552] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.228571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:16016 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.228577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.240447] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.240465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17236 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.240472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.251117] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.251136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:13828 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.251142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.264338] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.264356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:9992 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.264362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.277171] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.277188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:14543 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.277195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.290531] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.290550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:11210 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.290556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.303130] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.303148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:8663 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.303155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.316072] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.316090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:1037 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.316097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.328738] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.328756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:24772 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.328766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.341883] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.341900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:21176 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.341907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.732 [2024-11-26 19:26:02.352818] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.732 [2024-11-26 19:26:02.352835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:8491 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.732 [2024-11-26 19:26:02.352841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.994 [2024-11-26 19:26:02.364752] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.994 [2024-11-26 19:26:02.364769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:12251 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.994 [2024-11-26 19:26:02.364775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.994 [2024-11-26 19:26:02.377285] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.994 [2024-11-26 19:26:02.377302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:13505 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.994 [2024-11-26 19:26:02.377309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.994 [2024-11-26 19:26:02.391189] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.994 [2024-11-26 19:26:02.391207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:19548 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.994 [2024-11-26 19:26:02.391214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.994 [2024-11-26 19:26:02.403916] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.994 [2024-11-26 19:26:02.403935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:4233 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.994 [2024-11-26 19:26:02.403941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.994 [2024-11-26 19:26:02.416837] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.416854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:14504 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.416864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.428552] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.428569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:13575 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.428577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.442378] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.442400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:13923 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.442406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.454630] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.454647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:9579 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.454654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.466805] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.466822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:3352 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.466829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.480095] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.480113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:11719 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.480119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.492099] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.492117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:3563 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.492124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.504555] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.504572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20539 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.504579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.518046] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.518064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:579 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.518071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.530001] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.530018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:3365 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.530025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.543143] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.543161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:1409 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.543168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.553495] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.553512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:5715 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.553519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.566387] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.566405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:14764 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.566411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.578428] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.578446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:15665 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.578452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.591158] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.591176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:19235 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.591182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:49.995 [2024-11-26 19:26:02.605037] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:49.995 [2024-11-26 19:26:02.605055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19450 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:49.995 [2024-11-26 19:26:02.605062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.618244] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.618262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:18250 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.618268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:123 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.629445] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.629463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:16875 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.629470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.642597] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.642614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:6718 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.642621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.655487] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.655505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:15981 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.655514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.668559] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.668577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:4994 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.668583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.681089] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.681107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:19271 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.681113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.691299] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.691317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:15791 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.691323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.704902] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.704920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:12753 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.704927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.717540] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.717558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:9513 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.717565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.731770] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.731787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:804 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.731794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.743956] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.743973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:12742 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.743980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.755692] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.755710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:17220 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.755716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.769164] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.769181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:7045 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.769188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.781610] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.781627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:10070 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.781634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.794592] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.794610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:25149 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.794617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 19966.00 IOPS, 77.99 MiB/s [2024-11-26T18:26:02.883Z] [2024-11-26 19:26:02.808572] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.808589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:22221 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.808596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.822614] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.822633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:9292 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.822639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.837905] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.837923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:2973 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.837929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.258 [2024-11-26 19:26:02.851119] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.258 [2024-11-26 19:26:02.851136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:6212 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.258 [2024-11-26 19:26:02.851142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.259 [2024-11-26 19:26:02.861354] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.259 [2024-11-26 19:26:02.861372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:15719 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.259 [2024-11-26 19:26:02.861378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.259 [2024-11-26 19:26:02.874985] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.259 [2024-11-26 19:26:02.875003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:20898 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.259 [2024-11-26 19:26:02.875012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.888074] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.888093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:4011 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.888099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.901979] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.901996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:22648 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.902003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.914325] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.914343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:3833 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.914350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.928144] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.928161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3102 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.928168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.940660] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.940678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:11553 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.940684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.952645] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.952662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:20710 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.952673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.966362] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.966380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:17503 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.966387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.977176] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.977193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:4294 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.977199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.520 [2024-11-26 19:26:02.990144] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.520 [2024-11-26 19:26:02.990164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:7569 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.520 [2024-11-26 19:26:02.990171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.003800] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.003817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:22284 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.003824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.015089] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.015106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:22882 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.015113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.028550] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.028568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:6926 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.028574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.040036] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.040053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:7293 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.040060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.053834] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.053852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:14807 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.053858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.066687] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.066704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:25275 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.066711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.079940] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.079957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:13019 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.079964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.092242] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.092259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:3413 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.092266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.104389] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.104406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:2981 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.104413] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.115771] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.115789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:22547 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.115795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.128389] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.128406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:16005 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.128412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.521 [2024-11-26 19:26:03.141497] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.521 [2024-11-26 19:26:03.141514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:11644 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.521 [2024-11-26 19:26:03.141521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.783 [2024-11-26 19:26:03.153775] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.783 [2024-11-26 19:26:03.153793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:13378 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.783 [2024-11-26 19:26:03.153800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.783 [2024-11-26 19:26:03.165990] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.783 [2024-11-26 19:26:03.166007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:20996 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.783 [2024-11-26 19:26:03.166014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.783 [2024-11-26 19:26:03.179280] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.783 [2024-11-26 19:26:03.179298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:19663 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.783 [2024-11-26 19:26:03.179304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.783 [2024-11-26 19:26:03.191049] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.783 [2024-11-26 19:26:03.191066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:578 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.783 [2024-11-26 19:26:03.191073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.783 [2024-11-26 19:26:03.204056] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.783 [2024-11-26 19:26:03.204073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:13038 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.783 [2024-11-26 19:26:03.204083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.783 [2024-11-26 19:26:03.216671] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.216688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:20505 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.216695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.229824] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.229841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:9967 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.229847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.241457] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.241474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:5866 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.241481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.255267] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.255284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:16469 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.255291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.267232] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.267250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:4809 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.267256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.279253] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.279270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:2699 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.279277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.291904] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.291922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:16037 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.291928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.303619] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.303637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:8341 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.303644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.316381] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.316398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:8575 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.316404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.328252] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.328270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:953 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.328277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.341486] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.341504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:24213 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.341511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.354938] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.354956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:20795 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.354963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.367114] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.367131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:5267 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.367138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.378739] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.378756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:1791 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.378762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.391250] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.391268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:10185 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.391274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:50.784 [2024-11-26 19:26:03.404325] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:50.784 [2024-11-26 19:26:03.404342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:8004 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:50.784 [2024-11-26 19:26:03.404348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.417767] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.417785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:10065 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.417794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.428277] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.428294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:8016 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.428301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.440510] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.440528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:15685 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.440535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:74 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.453740] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.453757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:9503 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.453763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.466808] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.466825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:6641 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.466832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.480887] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.480905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3092 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.480911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.493425] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.493442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:5627 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.493449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.506227] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.506245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:3241 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.506252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.520232] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.520249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:14236 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.520256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.532675] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.532696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:22303 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.532702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.046 [2024-11-26 19:26:03.545522] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.046 [2024-11-26 19:26:03.545539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:10954 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.046 [2024-11-26 19:26:03.545545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.555405] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.555422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:19758 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.555429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.569622] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.569640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:10348 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.569647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.581966] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.581984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:21174 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.581991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.594787] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.594806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:20614 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.594812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.607085] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.607102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:20845 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.607109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.619460] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.619479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:19307 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.619486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.632212] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.632229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:1937 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.632236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.646223] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.646240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:7662 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.646247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.047 [2024-11-26 19:26:03.658697] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.047 [2024-11-26 19:26:03.658714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:15764 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.047 [2024-11-26 19:26:03.658721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.671868] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.671886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:1170 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.671892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.684457] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.684475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:16083 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.684482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.696559] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.696577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:1462 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.696583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.710758] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.710777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:1344 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.710784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.723400] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.723417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:11979 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.723424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.733653] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.733671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:5139 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.733678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.748110] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.748128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:10366 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.748138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.762105] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.762124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:8573 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.762131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.776042] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.776060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:15854 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.776068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 [2024-11-26 19:26:03.788758] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.788776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:14166 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.788782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 20016.50 IOPS, 78.19 MiB/s [2024-11-26T18:26:03.934Z] [2024-11-26 19:26:03.802417] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x134ae80) 00:29:51.309 [2024-11-26 19:26:03.802432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:6872 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:51.309 [2024-11-26 19:26:03.802438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:51.309 00:29:51.309 Latency(us) 00:29:51.309 [2024-11-26T18:26:03.934Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:51.309 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:29:51.309 nvme0n1 : 2.00 20046.08 78.31 0.00 0.00 6380.63 2007.04 19223.89 00:29:51.309 [2024-11-26T18:26:03.934Z] =================================================================================================================== 00:29:51.309 [2024-11-26T18:26:03.934Z] Total : 20046.08 78.31 0.00 0.00 6380.63 2007.04 19223.89 00:29:51.309 { 00:29:51.309 "results": [ 00:29:51.309 { 00:29:51.309 "job": "nvme0n1", 00:29:51.309 "core_mask": "0x2", 00:29:51.309 "workload": "randread", 00:29:51.309 "status": "finished", 00:29:51.309 "queue_depth": 128, 00:29:51.309 "io_size": 4096, 00:29:51.309 "runtime": 2.003434, 00:29:51.309 "iops": 20046.08087913053, 00:29:51.309 "mibps": 78.30500343410364, 00:29:51.309 "io_failed": 0, 00:29:51.309 "io_timeout": 0, 00:29:51.309 "avg_latency_us": 6380.630750562319, 00:29:51.309 "min_latency_us": 2007.04, 00:29:51.309 "max_latency_us": 19223.893333333333 00:29:51.309 } 00:29:51.309 ], 00:29:51.309 "core_count": 1 00:29:51.309 } 00:29:51.309 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:29:51.309 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:29:51.309 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:29:51.309 | .driver_specific 00:29:51.309 | .nvme_error 00:29:51.309 | .status_code 00:29:51.309 | .command_transient_transport_error' 00:29:51.309 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:29:51.571 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 157 > 0 )) 00:29:51.571 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 4048979 00:29:51.571 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 4048979 ']' 00:29:51.571 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 4048979 00:29:51.571 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:29:51.571 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:51.571 19:26:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4048979 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4048979' 00:29:51.571 killing process with pid 4048979 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 4048979 00:29:51.571 Received shutdown signal, test time was about 2.000000 seconds 00:29:51.571 00:29:51.571 Latency(us) 00:29:51.571 [2024-11-26T18:26:04.196Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:51.571 [2024-11-26T18:26:04.196Z] =================================================================================================================== 00:29:51.571 [2024-11-26T18:26:04.196Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 4048979 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@109 -- # run_bperf_err randread 131072 16 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=4049748 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 4049748 /var/tmp/bperf.sock 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 4049748 ']' 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:51.571 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:51.571 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:51.571 [2024-11-26 19:26:04.179535] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:51.571 [2024-11-26 19:26:04.179582] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4049748 ] 00:29:51.571 I/O size of 131072 is greater than zero copy threshold (65536). 00:29:51.571 Zero copy mechanism will not be used. 00:29:51.832 [2024-11-26 19:26:04.234359] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:51.832 [2024-11-26 19:26:04.263861] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:51.832 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:51.832 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:29:51.832 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:51.832 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:52.093 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:29:52.093 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.093 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:52.093 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.093 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:52.093 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:52.354 nvme0n1 00:29:52.354 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:29:52.354 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.354 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:52.354 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.354 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:29:52.354 19:26:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:52.354 I/O size of 131072 is greater than zero copy threshold (65536). 00:29:52.354 Zero copy mechanism will not be used. 00:29:52.354 Running I/O for 2 seconds... 00:29:52.354 [2024-11-26 19:26:04.824843] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.354 [2024-11-26 19:26:04.824881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:22144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.354 [2024-11-26 19:26:04.824891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.354 [2024-11-26 19:26:04.833689] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.354 [2024-11-26 19:26:04.833711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:11328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.354 [2024-11-26 19:26:04.833719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.354 [2024-11-26 19:26:04.844442] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.354 [2024-11-26 19:26:04.844464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.354 [2024-11-26 19:26:04.844471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.354 [2024-11-26 19:26:04.852671] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.852691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:10208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.852698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.859872] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.859891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.859898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.865210] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.865229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.865235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.872938] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.872957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:15744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.872964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.882592] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.882611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:7744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.882618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.890240] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.890259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:13152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.890265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.897140] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.897158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.897165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.905800] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.905819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:5408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.905826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.916020] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.916040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.916052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.928722] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.928742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:14048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.928748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.940719] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.940738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.940745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.950582] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.950601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:12896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.950607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.958477] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.958496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:3392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.958502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.967260] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.967278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.967285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.355 [2024-11-26 19:26:04.973403] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.355 [2024-11-26 19:26:04.973422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.355 [2024-11-26 19:26:04.973429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:04.977639] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:04.977658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:18528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:04.977665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:04.982962] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:04.982981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:04.982988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:04.992442] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:04.992464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:04.992471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.000769] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.000787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:7776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.000794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.013058] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.013077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.013084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.025402] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.025421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:11616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.025428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.035616] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.035634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:1216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.035641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.041253] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.041271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:13664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.041278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.048476] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.048495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:7648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.048501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.058847] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.058871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.058877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.065283] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.065302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:21280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.065308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.071071] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.071089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.071095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.078708] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.078727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.078733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.086153] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.086171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.086179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.095140] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.617 [2024-11-26 19:26:05.095158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:4576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.617 [2024-11-26 19:26:05.095165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.617 [2024-11-26 19:26:05.101860] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.101883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:22368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.101890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.112652] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.112670] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:13024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.112676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.123994] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.124013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:21056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.124019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.131496] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.131514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.131520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.137054] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.137072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.137082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.144990] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.145009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.145015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.148419] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.148437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:16256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.148444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.158832] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.158850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:22240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.158857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.167999] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.168017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.168023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.176221] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.176239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:4288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.176246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.184049] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.184067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.184073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.192546] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.192564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:6976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.192570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.200923] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.200940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:10176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.200946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.206134] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.206155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.206162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.214485] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.214506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.214512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.224909] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.224927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.224934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.618 [2024-11-26 19:26:05.235263] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.618 [2024-11-26 19:26:05.235280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.618 [2024-11-26 19:26:05.235287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.879 [2024-11-26 19:26:05.244054] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.879 [2024-11-26 19:26:05.244072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.879 [2024-11-26 19:26:05.244078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.255393] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.255411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.255417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.263119] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.263137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.263143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.272612] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.272629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.272636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.282005] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.282022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.282029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.294136] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.294154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.294160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.304196] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.304213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.304219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.312337] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.312355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.312361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.323500] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.323519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.323525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.334832] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.334850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.334856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.347743] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.347760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.347767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.359577] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.359595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.359601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.370233] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.370250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.370256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.379720] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.379741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.379748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.389132] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.389150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.389156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.399104] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.399122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.399129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.409119] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.409137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.409144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.419333] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.419351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.419358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.429347] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.429365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.429371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.441246] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.441264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.441270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.451924] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.451942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.451948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.462536] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.462553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.462560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.473928] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.473947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.473953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.483077] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.483095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:5792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.483101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:52.880 [2024-11-26 19:26:05.492611] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:52.880 [2024-11-26 19:26:05.492628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:13696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:52.880 [2024-11-26 19:26:05.492635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.142 [2024-11-26 19:26:05.504121] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.142 [2024-11-26 19:26:05.504140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.142 [2024-11-26 19:26:05.504146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.142 [2024-11-26 19:26:05.515735] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.142 [2024-11-26 19:26:05.515754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:4800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.142 [2024-11-26 19:26:05.515760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.142 [2024-11-26 19:26:05.528046] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.142 [2024-11-26 19:26:05.528065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:18240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.142 [2024-11-26 19:26:05.528073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.142 [2024-11-26 19:26:05.540593] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.142 [2024-11-26 19:26:05.540611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.142 [2024-11-26 19:26:05.540618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.553685] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.553703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:10112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.553710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.564845] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.564868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:20160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.564878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.576220] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.576238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.576244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.586112] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.586129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.586135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.596387] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.596405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.596411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.608279] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.608297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.608303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.616363] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.616381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.616388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.625905] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.625923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.625930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.637319] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.637337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.637344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.647174] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.647193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.647199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.655548] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.655570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.655577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.665099] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.665118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:24352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.665125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.676077] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.676096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.676103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.686926] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.686944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.686951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.697208] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.697227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.697233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.707897] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.707916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.707923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.718188] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.718207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:2688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.718213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.726889] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.726908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.726914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.738225] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.738244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.738250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.749203] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.749222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:7808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.749228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.143 [2024-11-26 19:26:05.759735] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.143 [2024-11-26 19:26:05.759753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:7904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.143 [2024-11-26 19:26:05.759760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.405 [2024-11-26 19:26:05.771814] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.405 [2024-11-26 19:26:05.771834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:22336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.405 [2024-11-26 19:26:05.771840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.405 [2024-11-26 19:26:05.784452] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.405 [2024-11-26 19:26:05.784471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:3008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.405 [2024-11-26 19:26:05.784477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.405 [2024-11-26 19:26:05.796143] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.796163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.796169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.808508] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.808528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:9632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.808535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.406 3217.00 IOPS, 402.12 MiB/s [2024-11-26T18:26:06.031Z] [2024-11-26 19:26:05.821122] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.821141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:17952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.821148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.833979] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.833997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.834004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.846886] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.846908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:24320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.846915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.859202] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.859221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:8992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.859228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.870669] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.870688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:20736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.870695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.879704] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.879723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.879730] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.889475] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.889494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.889502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.900219] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.900238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:9824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.900244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.911380] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.911400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.911406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.921302] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.921320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.921327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.932055] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.932074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:3584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.932080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.941665] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.941684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.941690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.953103] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.953122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.953129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.965998] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.966017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:10336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.966024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.978455] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.978474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.978480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:05.991647] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:05.991666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:3712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:05.991673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:06.003969] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:06.003989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:06.003995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:06.016362] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:06.016381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:24512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:06.016388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.406 [2024-11-26 19:26:06.027931] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.406 [2024-11-26 19:26:06.027949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:14592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.406 [2024-11-26 19:26:06.027956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.668 [2024-11-26 19:26:06.040120] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.668 [2024-11-26 19:26:06.040140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.668 [2024-11-26 19:26:06.040151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.668 [2024-11-26 19:26:06.052331] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.668 [2024-11-26 19:26:06.052349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:7040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.668 [2024-11-26 19:26:06.052356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.668 [2024-11-26 19:26:06.064419] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.668 [2024-11-26 19:26:06.064438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:6272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.668 [2024-11-26 19:26:06.064445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.077169] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.077189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:1600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.077196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.089594] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.089613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:4960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.089620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.100757] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.100776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:8224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.100783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.112279] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.112298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:4128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.112304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.123797] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.123816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:18848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.123822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.133813] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.133832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.133839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.141696] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.141718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:13216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.141725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.150313] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.150333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:25216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.150339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.159698] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.159717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:24640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.159724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.169044] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.169063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:14976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.169069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.178367] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.178387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.178393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.189586] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.189605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:14624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.189611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.199828] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.199847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:14688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.199854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.210317] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.210336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:17472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.210343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.220885] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.220904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.220910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.231946] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.231965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:6112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.231972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.240602] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.240621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:7680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.240629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.251379] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.251399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:19744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.251405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.263772] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.263791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:21344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.263798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.274167] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.274185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.274192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.669 [2024-11-26 19:26:06.284758] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.669 [2024-11-26 19:26:06.284776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:21888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.669 [2024-11-26 19:26:06.284783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.296444] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.296463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.296470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.307953] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.307971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.307978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.318779] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.318798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:25120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.318808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.328218] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.328237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:20768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.328243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.338361] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.338379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.338385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.349041] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.349060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:5440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.349067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.359997] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.360015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:20224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.360021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.370731] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.370750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:21472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.370757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.382459] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.382477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:18368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.382484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.393693] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.393711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:2080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.393717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.400131] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.400148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:5792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.400154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.411832] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.411851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:15104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.411858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.932 [2024-11-26 19:26:06.421624] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.932 [2024-11-26 19:26:06.421642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:21536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.932 [2024-11-26 19:26:06.421649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.430774] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.430794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.430802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.442517] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.442536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.442542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.452240] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.452258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:7616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.452265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.463332] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.463351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.463358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.473923] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.473941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:5824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.473948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.484905] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.484924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.484931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.497247] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.497265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:12608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.497275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.509189] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.509208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.509215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.519791] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.519810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:10496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.519816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.530600] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.530619] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.530626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.540829] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.540848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:20640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.540854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:53.933 [2024-11-26 19:26:06.551204] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:53.933 [2024-11-26 19:26:06.551222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:2368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:53.933 [2024-11-26 19:26:06.551229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.561913] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.561932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.561938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.571328] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.571347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.571353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.578454] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.578473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:9536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.578479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.589721] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.589743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.589750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.599017] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.599035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.599042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.608384] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.608403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:2016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.608410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.618770] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.618790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:22048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.618796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.629424] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.629443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.629450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.640115] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.640134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.640141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.651054] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.651072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.651079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.662345] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.662365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:15488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.662371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.670484] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.194 [2024-11-26 19:26:06.670503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.194 [2024-11-26 19:26:06.670509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:54.194 [2024-11-26 19:26:06.679752] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.679770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:2432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.679777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.688953] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.688972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:32 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.688979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.699941] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.699959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:19328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.699966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.711604] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.711623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:9376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.711629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.720757] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.720776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.720782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.730255] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.730273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.730280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.740622] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.740640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.740646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.751062] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.751082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:23712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.751088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.761331] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.761350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:1344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.761360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.769122] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.769140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.769147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.775936] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.775954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:5792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.775961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.786109] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.786127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:8416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.786133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.795119] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.795138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.795145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.806106] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.806124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:18880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.806131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:54.195 [2024-11-26 19:26:06.816093] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x23b0c10) 00:29:54.195 [2024-11-26 19:26:06.816112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:5888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:54.195 [2024-11-26 19:26:06.816119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:54.457 3072.50 IOPS, 384.06 MiB/s 00:29:54.457 Latency(us) 00:29:54.457 [2024-11-26T18:26:07.082Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:54.457 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:29:54.457 nvme0n1 : 2.00 3076.15 384.52 0.00 0.00 5197.68 914.77 14090.24 00:29:54.457 [2024-11-26T18:26:07.082Z] =================================================================================================================== 00:29:54.457 [2024-11-26T18:26:07.082Z] Total : 3076.15 384.52 0.00 0.00 5197.68 914.77 14090.24 00:29:54.457 { 00:29:54.457 "results": [ 00:29:54.457 { 00:29:54.457 "job": "nvme0n1", 00:29:54.457 "core_mask": "0x2", 00:29:54.457 "workload": "randread", 00:29:54.457 "status": "finished", 00:29:54.457 "queue_depth": 16, 00:29:54.457 "io_size": 131072, 00:29:54.457 "runtime": 2.002829, 00:29:54.457 "iops": 3076.148787540025, 00:29:54.457 "mibps": 384.5185984425031, 00:29:54.457 "io_failed": 0, 00:29:54.457 "io_timeout": 0, 00:29:54.457 "avg_latency_us": 5197.681726992371, 00:29:54.457 "min_latency_us": 914.7733333333333, 00:29:54.457 "max_latency_us": 14090.24 00:29:54.457 } 00:29:54.457 ], 00:29:54.457 "core_count": 1 00:29:54.457 } 00:29:54.457 19:26:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:29:54.457 19:26:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:29:54.457 19:26:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:29:54.457 | .driver_specific 00:29:54.457 | .nvme_error 00:29:54.457 | .status_code 00:29:54.457 | .command_transient_transport_error' 00:29:54.457 19:26:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:29:54.457 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 199 > 0 )) 00:29:54.457 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 4049748 00:29:54.457 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 4049748 ']' 00:29:54.457 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 4049748 00:29:54.457 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:29:54.457 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:54.457 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4049748 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4049748' 00:29:54.718 killing process with pid 4049748 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 4049748 00:29:54.718 Received shutdown signal, test time was about 2.000000 seconds 00:29:54.718 00:29:54.718 Latency(us) 00:29:54.718 [2024-11-26T18:26:07.343Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:54.718 [2024-11-26T18:26:07.343Z] =================================================================================================================== 00:29:54.718 [2024-11-26T18:26:07.343Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 4049748 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@114 -- # run_bperf_err randwrite 4096 128 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=4050254 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 4050254 /var/tmp/bperf.sock 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 4050254 ']' 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:54.718 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:54.718 19:26:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:54.718 [2024-11-26 19:26:07.238756] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:54.718 [2024-11-26 19:26:07.238815] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4050254 ] 00:29:54.718 [2024-11-26 19:26:07.329764] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:54.985 [2024-11-26 19:26:07.359638] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:55.560 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:55.560 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:29:55.560 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:55.560 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:55.821 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:29:55.821 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.821 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:55.821 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.821 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:55.821 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:56.083 nvme0n1 00:29:56.083 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:29:56.083 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.083 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:56.083 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.083 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:29:56.083 19:26:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:56.083 Running I/O for 2 seconds... 00:29:56.083 [2024-11-26 19:26:08.694396] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef92c0 00:29:56.083 [2024-11-26 19:26:08.695367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:6500 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.083 [2024-11-26 19:26:08.695397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.706560] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef9b30 00:29:56.344 [2024-11-26 19:26:08.707488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:23778 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.707515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.718469] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efa3a0 00:29:56.344 [2024-11-26 19:26:08.719373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:6908 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.719390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.730368] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efac10 00:29:56.344 [2024-11-26 19:26:08.731248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:2118 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.731265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.742255] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efb480 00:29:56.344 [2024-11-26 19:26:08.743113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:6 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.743130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.754124] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efbcf0 00:29:56.344 [2024-11-26 19:26:08.754961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:17990 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.754978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.765992] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efc560 00:29:56.344 [2024-11-26 19:26:08.766805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:8652 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.766821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.777857] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efcdd0 00:29:56.344 [2024-11-26 19:26:08.778651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:20587 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.778667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.789714] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efd640 00:29:56.344 [2024-11-26 19:26:08.790489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:5874 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.344 [2024-11-26 19:26:08.790505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:29:56.344 [2024-11-26 19:26:08.801607] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efdeb0 00:29:56.345 [2024-11-26 19:26:08.802360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:16689 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.802377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.813475] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efe720 00:29:56.345 [2024-11-26 19:26:08.814212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:22885 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.814229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.825348] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eff3c8 00:29:56.345 [2024-11-26 19:26:08.826067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:3445 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.826083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.839620] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eff3c8 00:29:56.345 [2024-11-26 19:26:08.840989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:21328 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.841006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.851483] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efe720 00:29:56.345 [2024-11-26 19:26:08.852820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:2279 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.852841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.863352] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efdeb0 00:29:56.345 [2024-11-26 19:26:08.864668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:16133 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.864684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.875201] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efd640 00:29:56.345 [2024-11-26 19:26:08.876496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:6524 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.876513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.887051] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efcdd0 00:29:56.345 [2024-11-26 19:26:08.888327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:1527 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.888343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.898912] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efc560 00:29:56.345 [2024-11-26 19:26:08.900165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:12764 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.900182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.910767] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efbcf0 00:29:56.345 [2024-11-26 19:26:08.912009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:5789 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.912025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.922669] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efb480 00:29:56.345 [2024-11-26 19:26:08.923891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:21199 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.923908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.934522] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efac10 00:29:56.345 [2024-11-26 19:26:08.935717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:9582 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.935733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.946368] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efa3a0 00:29:56.345 [2024-11-26 19:26:08.947543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:8115 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.947561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:29:56.345 [2024-11-26 19:26:08.960638] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee27f0 00:29:56.345 [2024-11-26 19:26:08.962463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:17611 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.345 [2024-11-26 19:26:08.962480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:29:56.606 [2024-11-26 19:26:08.972524] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee3060 00:29:56.606 [2024-11-26 19:26:08.974328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:18919 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.606 [2024-11-26 19:26:08.974345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:29:56.606 [2024-11-26 19:26:08.984392] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee38d0 00:29:56.606 [2024-11-26 19:26:08.986175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:22510 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.606 [2024-11-26 19:26:08.986191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:29:56.606 [2024-11-26 19:26:08.996267] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee4140 00:29:56.606 [2024-11-26 19:26:08.998027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:20287 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.606 [2024-11-26 19:26:08.998044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:29:56.606 [2024-11-26 19:26:09.008134] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee49b0 00:29:56.606 [2024-11-26 19:26:09.009874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:4319 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.009891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.019997] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee5220 00:29:56.607 [2024-11-26 19:26:09.021711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:10096 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.021731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.031848] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee5a90 00:29:56.607 [2024-11-26 19:26:09.033556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:23902 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.033573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.043723] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee6300 00:29:56.607 [2024-11-26 19:26:09.045403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:15492 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.045420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.054079] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee6b70 00:29:56.607 [2024-11-26 19:26:09.055099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:15064 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.055115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.065212] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee73e0 00:29:56.607 [2024-11-26 19:26:09.066210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:15943 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.066227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.077075] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee7c50 00:29:56.607 [2024-11-26 19:26:09.078055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:3413 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.078071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.088936] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee84c0 00:29:56.607 [2024-11-26 19:26:09.089897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:9399 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.089914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.103227] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef4f40 00:29:56.607 [2024-11-26 19:26:09.104832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20838 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.104849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.115113] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef46d0 00:29:56.607 [2024-11-26 19:26:09.116697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:5494 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.116717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.126978] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef3e60 00:29:56.607 [2024-11-26 19:26:09.128551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:13595 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.128568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.138851] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef35f0 00:29:56.607 [2024-11-26 19:26:09.140399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:10298 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.140416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.148506] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efa7d8 00:29:56.607 [2024-11-26 19:26:09.149396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:6122 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.149412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.161283] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee2c28 00:29:56.607 [2024-11-26 19:26:09.162182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:12857 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.162199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.172428] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee23b8 00:29:56.607 [2024-11-26 19:26:09.173306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:19879 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.173323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.184306] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee1b48 00:29:56.607 [2024-11-26 19:26:09.185163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:18592 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.185179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.196184] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee12d8 00:29:56.607 [2024-11-26 19:26:09.197022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:1136 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.197039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.208060] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0a68 00:29:56.607 [2024-11-26 19:26:09.208876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:1532 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.208893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:29:56.607 [2024-11-26 19:26:09.219931] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee01f8 00:29:56.607 [2024-11-26 19:26:09.220725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:24993 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.607 [2024-11-26 19:26:09.220741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.234231] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efc998 00:29:56.869 [2024-11-26 19:26:09.235677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:1710 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.235697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.246093] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efd208 00:29:56.869 [2024-11-26 19:26:09.247514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:22972 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.247535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.257964] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efda78 00:29:56.869 [2024-11-26 19:26:09.259363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:18521 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.259379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.269818] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efe2e8 00:29:56.869 [2024-11-26 19:26:09.271205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:16224 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.271221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.281691] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efeb58 00:29:56.869 [2024-11-26 19:26:09.283057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:22613 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.283074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.295943] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ede8a8 00:29:56.869 [2024-11-26 19:26:09.297956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23866 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.297973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.307807] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ede038 00:29:56.869 [2024-11-26 19:26:09.309801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:15645 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.309818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.319665] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eef270 00:29:56.869 [2024-11-26 19:26:09.321632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:1772 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.321648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.330031] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eefae0 00:29:56.869 [2024-11-26 19:26:09.331340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:1724 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.331359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.341205] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef0350 00:29:56.869 [2024-11-26 19:26:09.342494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:16649 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.342511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:52 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.355474] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edf988 00:29:56.869 [2024-11-26 19:26:09.357414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:21058 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.357431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.367344] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee01f8 00:29:56.869 [2024-11-26 19:26:09.369261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:12284 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.869 [2024-11-26 19:26:09.369277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:29:56.869 [2024-11-26 19:26:09.379211] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0a68 00:29:56.869 [2024-11-26 19:26:09.381108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:52 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.381125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.391080] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee12d8 00:29:56.870 [2024-11-26 19:26:09.392955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:7675 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.392972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.402943] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee1b48 00:29:56.870 [2024-11-26 19:26:09.404791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:7669 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.404807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.414793] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee23b8 00:29:56.870 [2024-11-26 19:26:09.416631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:20473 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.416648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.426659] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee2c28 00:29:56.870 [2024-11-26 19:26:09.428469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8769 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.428486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.438523] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee3498 00:29:56.870 [2024-11-26 19:26:09.440328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:14736 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.440346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.450385] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee3d08 00:29:56.870 [2024-11-26 19:26:09.452158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:7314 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.452175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.460042] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee84c0 00:29:56.870 [2024-11-26 19:26:09.461159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:2040 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.461177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.474337] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef46d0 00:29:56.870 [2024-11-26 19:26:09.476100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:11159 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.476116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:29:56.870 [2024-11-26 19:26:09.486279] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef4f40 00:29:56.870 [2024-11-26 19:26:09.488021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:6133 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.870 [2024-11-26 19:26:09.488037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.498158] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef57b0 00:29:57.132 [2024-11-26 19:26:09.499878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:18082 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.499895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.510012] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef6020 00:29:57.132 [2024-11-26 19:26:09.511709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:706 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.511726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.521870] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef6890 00:29:57.132 [2024-11-26 19:26:09.523550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:3609 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.523566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.531520] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef6458 00:29:57.132 [2024-11-26 19:26:09.532540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:4618 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.532557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.545806] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee6738 00:29:57.132 [2024-11-26 19:26:09.547477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:24290 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.547494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.557666] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee6fa8 00:29:57.132 [2024-11-26 19:26:09.559313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:17103 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.559330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.569558] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee7818 00:29:57.132 [2024-11-26 19:26:09.571187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:24794 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.571204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.581415] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee8088 00:29:57.132 [2024-11-26 19:26:09.583025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:7007 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.583041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.593270] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee88f8 00:29:57.132 [2024-11-26 19:26:09.594851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:9279 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.594871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.605136] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee9168 00:29:57.132 [2024-11-26 19:26:09.606698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:19525 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.606714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.615505] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee99d8 00:29:57.132 [2024-11-26 19:26:09.616413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:1567 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.616430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:59 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.626655] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eea248 00:29:57.132 [2024-11-26 19:26:09.627545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:13365 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.627562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.639422] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efa3a0 00:29:57.132 [2024-11-26 19:26:09.640320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:14290 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.640340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.650570] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efac10 00:29:57.132 [2024-11-26 19:26:09.651449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:4861 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.651465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.664822] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee27f0 00:29:57.132 [2024-11-26 19:26:09.666346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:11611 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.666363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.676712] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee1f80 00:29:57.132 [2024-11-26 19:26:09.678220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:3960 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.678237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.687054] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee1710 00:29:57.132 21346.00 IOPS, 83.38 MiB/s [2024-11-26T18:26:09.757Z] [2024-11-26 19:26:09.687887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:19175 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.687903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:59 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.698457] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0ea0 00:29:57.132 [2024-11-26 19:26:09.699275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:24701 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.699292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.710328] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0630 00:29:57.132 [2024-11-26 19:26:09.711123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23856 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.711140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.724580] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef0bc0 00:29:57.132 [2024-11-26 19:26:09.726031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:22731 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.726051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.736466] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef0350 00:29:57.132 [2024-11-26 19:26:09.737889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:2297 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.132 [2024-11-26 19:26:09.737906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:29:57.132 [2024-11-26 19:26:09.748327] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eefae0 00:29:57.132 [2024-11-26 19:26:09.749732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:5630 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.133 [2024-11-26 19:26:09.749748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:29:57.394 [2024-11-26 19:26:09.760186] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eef270 00:29:57.394 [2024-11-26 19:26:09.761563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:19827 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.394 [2024-11-26 19:26:09.761579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:29:57.394 [2024-11-26 19:26:09.772025] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eff3c8 00:29:57.394 [2024-11-26 19:26:09.773385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:6780 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.394 [2024-11-26 19:26:09.773402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:29:57.394 [2024-11-26 19:26:09.786257] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eee190 00:29:57.394 [2024-11-26 19:26:09.788263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:6581 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.394 [2024-11-26 19:26:09.788279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:29:57.394 [2024-11-26 19:26:09.798119] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eeea00 00:29:57.394 [2024-11-26 19:26:09.800108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:20070 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.394 [2024-11-26 19:26:09.800124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:29:57.394 [2024-11-26 19:26:09.809980] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efe2e8 00:29:57.394 [2024-11-26 19:26:09.811944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:2345 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.394 [2024-11-26 19:26:09.811961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:29:57.394 [2024-11-26 19:26:09.821828] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efda78 00:29:57.394 [2024-11-26 19:26:09.823774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:906 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.823790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.833684] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efd208 00:29:57.395 [2024-11-26 19:26:09.835611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:6904 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.835627] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.845542] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efc998 00:29:57.395 [2024-11-26 19:26:09.847453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:5198 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.847470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.857413] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efc128 00:29:57.395 [2024-11-26 19:26:09.859301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:9345 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.859317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.869270] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efb8b8 00:29:57.395 [2024-11-26 19:26:09.871136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:16805 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.871152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.881123] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efb048 00:29:57.395 [2024-11-26 19:26:09.882967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:11977 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.882983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.892977] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efa7d8 00:29:57.395 [2024-11-26 19:26:09.894795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:14333 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.894812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.904820] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef9f68 00:29:57.395 [2024-11-26 19:26:09.906620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:1762 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.906636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.916654] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef96f8 00:29:57.395 [2024-11-26 19:26:09.918438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:15149 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.918454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.928520] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef8e88 00:29:57.395 [2024-11-26 19:26:09.930285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:6503 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.930302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.938174] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef4f40 00:29:57.395 [2024-11-26 19:26:09.939274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:15722 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.939290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.950023] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef57b0 00:29:57.395 [2024-11-26 19:26:09.951104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:9439 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.951124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.961873] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef6020 00:29:57.395 [2024-11-26 19:26:09.962938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:17713 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.962955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.973717] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef6890 00:29:57.395 [2024-11-26 19:26:09.974757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:4924 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.974773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.985571] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef7100 00:29:57.395 [2024-11-26 19:26:09.986594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:15932 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:09.986610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:09.999828] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee6738 00:29:57.395 [2024-11-26 19:26:10.001503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:12860 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:10.001521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:29:57.395 [2024-11-26 19:26:10.012210] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee5ec8 00:29:57.395 [2024-11-26 19:26:10.013858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:14705 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.395 [2024-11-26 19:26:10.013881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.024102] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee5658 00:29:57.658 [2024-11-26 19:26:10.025736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:17124 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.025754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.035995] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee4de8 00:29:57.658 [2024-11-26 19:26:10.037605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:12237 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.037621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.047892] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee4578 00:29:57.658 [2024-11-26 19:26:10.049476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24555 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.049493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.059775] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee3d08 00:29:57.658 [2024-11-26 19:26:10.061352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:17498 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.061368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.071651] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee3498 00:29:57.658 [2024-11-26 19:26:10.073196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:14122 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.073212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.083495] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee2c28 00:29:57.658 [2024-11-26 19:26:10.085025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:15649 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.085042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.095363] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee23b8 00:29:57.658 [2024-11-26 19:26:10.096870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:16911 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.096886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.107209] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee1b48 00:29:57.658 [2024-11-26 19:26:10.108690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:21264 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.108707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.119059] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee12d8 00:29:57.658 [2024-11-26 19:26:10.120518] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:14602 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.120535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.130920] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0a68 00:29:57.658 [2024-11-26 19:26:10.132360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:3314 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.132377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.142765] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee01f8 00:29:57.658 [2024-11-26 19:26:10.144187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:3583 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.144204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.154619] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edf988 00:29:57.658 [2024-11-26 19:26:10.156023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:3422 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.156039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:94 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.166462] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edf118 00:29:57.658 [2024-11-26 19:26:10.167838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:433 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.167854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.178305] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eee5c8 00:29:57.658 [2024-11-26 19:26:10.179670] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:6055 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.658 [2024-11-26 19:26:10.179686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:29:57.658 [2024-11-26 19:26:10.190165] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eeee38 00:29:57.658 [2024-11-26 19:26:10.191508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:8251 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.191524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:29:57.659 [2024-11-26 19:26:10.202061] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efdeb0 00:29:57.659 [2024-11-26 19:26:10.203380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:16328 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.203397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:29:57.659 [2024-11-26 19:26:10.213911] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efd640 00:29:57.659 [2024-11-26 19:26:10.215205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:12407 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.215227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:29:57.659 [2024-11-26 19:26:10.225761] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efcdd0 00:29:57.659 [2024-11-26 19:26:10.227042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:15703 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.227062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:29:57.659 [2024-11-26 19:26:10.237632] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efc560 00:29:57.659 [2024-11-26 19:26:10.238891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:3703 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.238907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:29:57.659 [2024-11-26 19:26:10.249474] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efbcf0 00:29:57.659 [2024-11-26 19:26:10.250710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:6521 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.250727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:29:57.659 [2024-11-26 19:26:10.261323] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efb480 00:29:57.659 [2024-11-26 19:26:10.262539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:20715 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.262563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:29:57.659 [2024-11-26 19:26:10.273179] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efac10 00:29:57.659 [2024-11-26 19:26:10.274377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:10186 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.659 [2024-11-26 19:26:10.274393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.287429] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee23b8 00:29:57.921 [2024-11-26 19:26:10.289275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:1432 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.289292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.297077] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eea680 00:29:57.921 [2024-11-26 19:26:10.298260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:20051 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.298276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.308936] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee9e10 00:29:57.921 [2024-11-26 19:26:10.310097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:21701 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.310114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.320769] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee95a0 00:29:57.921 [2024-11-26 19:26:10.321920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:10360 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.321936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.332628] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee8d30 00:29:57.921 [2024-11-26 19:26:10.333750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:13233 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.333766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.344475] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee84c0 00:29:57.921 [2024-11-26 19:26:10.345577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:21733 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.345593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.358724] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef4b08 00:29:57.921 [2024-11-26 19:26:10.360480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12788 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.360495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.370561] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef5378 00:29:57.921 [2024-11-26 19:26:10.372298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:17031 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.372313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.382422] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef5be8 00:29:57.921 [2024-11-26 19:26:10.384135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:18947 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.384152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.394307] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef6458 00:29:57.921 [2024-11-26 19:26:10.396000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:5231 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.396017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.406175] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef6cc8 00:29:57.921 [2024-11-26 19:26:10.407840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:21650 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.407856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.418027] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef7538 00:29:57.921 [2024-11-26 19:26:10.419672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:17615 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.419689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.429883] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef7da8 00:29:57.921 [2024-11-26 19:26:10.431508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:20230 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.431524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.441738] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef8618 00:29:57.921 [2024-11-26 19:26:10.443346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:736 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.443362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.453582] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef8e88 00:29:57.921 [2024-11-26 19:26:10.455170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:16618 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.455186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.465435] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef96f8 00:29:57.921 [2024-11-26 19:26:10.467002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:4056 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.467018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.477306] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef9f68 00:29:57.921 [2024-11-26 19:26:10.478845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:21594 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.921 [2024-11-26 19:26:10.478861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:29:57.921 [2024-11-26 19:26:10.489234] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efa7d8 00:29:57.922 [2024-11-26 19:26:10.490760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:20183 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.922 [2024-11-26 19:26:10.490776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:52 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:29:57.922 [2024-11-26 19:26:10.501085] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016efb048 00:29:57.922 [2024-11-26 19:26:10.502585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:24952 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.922 [2024-11-26 19:26:10.502602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:29:57.922 [2024-11-26 19:26:10.510740] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef1ca0 00:29:57.922 [2024-11-26 19:26:10.511583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:17369 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.922 [2024-11-26 19:26:10.511599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:29:57.922 [2024-11-26 19:26:10.522585] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef1430 00:29:57.922 [2024-11-26 19:26:10.523408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22740 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.922 [2024-11-26 19:26:10.523425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:29:57.922 [2024-11-26 19:26:10.534447] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef0bc0 00:29:57.922 [2024-11-26 19:26:10.535253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:11685 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:57.922 [2024-11-26 19:26:10.535269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.546309] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ef0350 00:29:58.184 [2024-11-26 19:26:10.547093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:10471 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.547109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.560554] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0630 00:29:58.184 [2024-11-26 19:26:10.561993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:14010 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.562010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.572414] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edfdc0 00:29:58.184 [2024-11-26 19:26:10.573823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6776 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.573843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.584284] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edf550 00:29:58.184 [2024-11-26 19:26:10.585673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:9544 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.585690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:74 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.596145] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edece0 00:29:58.184 [2024-11-26 19:26:10.597511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:14355 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.597528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.607991] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eff3c8 00:29:58.184 [2024-11-26 19:26:10.609337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:3603 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.609353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.619835] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eef270 00:29:58.184 [2024-11-26 19:26:10.621166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:19576 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.621181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:29:58.184 [2024-11-26 19:26:10.631671] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016eefae0 00:29:58.184 [2024-11-26 19:26:10.632981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:20079 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.184 [2024-11-26 19:26:10.632997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:29:58.185 [2024-11-26 19:26:10.645946] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edf550 00:29:58.185 [2024-11-26 19:26:10.647903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:12295 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.185 [2024-11-26 19:26:10.647920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:29:58.185 [2024-11-26 19:26:10.657812] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016edfdc0 00:29:58.185 [2024-11-26 19:26:10.659749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:8491 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.185 [2024-11-26 19:26:10.659765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:29:58.185 [2024-11-26 19:26:10.669668] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0630 00:29:58.185 [2024-11-26 19:26:10.671585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:23864 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.185 [2024-11-26 19:26:10.671602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:29:58.185 [2024-11-26 19:26:10.681527] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9e9d0) with pdu=0x200016ee0ea0 00:29:58.185 [2024-11-26 19:26:10.683430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:20092 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:58.185 [2024-11-26 19:26:10.683449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:29:58.185 21338.50 IOPS, 83.35 MiB/s 00:29:58.185 Latency(us) 00:29:58.185 [2024-11-26T18:26:10.810Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:58.185 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:29:58.185 nvme0n1 : 2.00 21369.61 83.48 0.00 0.00 5985.15 1966.08 16930.13 00:29:58.185 [2024-11-26T18:26:10.810Z] =================================================================================================================== 00:29:58.185 [2024-11-26T18:26:10.810Z] Total : 21369.61 83.48 0.00 0.00 5985.15 1966.08 16930.13 00:29:58.185 { 00:29:58.185 "results": [ 00:29:58.185 { 00:29:58.185 "job": "nvme0n1", 00:29:58.185 "core_mask": "0x2", 00:29:58.185 "workload": "randwrite", 00:29:58.185 "status": "finished", 00:29:58.185 "queue_depth": 128, 00:29:58.185 "io_size": 4096, 00:29:58.185 "runtime": 2.003078, 00:29:58.185 "iops": 21369.61216687518, 00:29:58.185 "mibps": 83.47504752685617, 00:29:58.185 "io_failed": 0, 00:29:58.185 "io_timeout": 0, 00:29:58.185 "avg_latency_us": 5985.152931978352, 00:29:58.185 "min_latency_us": 1966.08, 00:29:58.185 "max_latency_us": 16930.133333333335 00:29:58.185 } 00:29:58.185 ], 00:29:58.185 "core_count": 1 00:29:58.185 } 00:29:58.185 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:29:58.185 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:29:58.185 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:29:58.185 | .driver_specific 00:29:58.185 | .nvme_error 00:29:58.185 | .status_code 00:29:58.185 | .command_transient_transport_error' 00:29:58.185 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 167 > 0 )) 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 4050254 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 4050254 ']' 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 4050254 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4050254 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4050254' 00:29:58.447 killing process with pid 4050254 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 4050254 00:29:58.447 Received shutdown signal, test time was about 2.000000 seconds 00:29:58.447 00:29:58.447 Latency(us) 00:29:58.447 [2024-11-26T18:26:11.072Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:58.447 [2024-11-26T18:26:11.072Z] =================================================================================================================== 00:29:58.447 [2024-11-26T18:26:11.072Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:58.447 19:26:10 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 4050254 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@115 -- # run_bperf_err randwrite 131072 16 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=4051062 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 4051062 /var/tmp/bperf.sock 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 4051062 ']' 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:58.447 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:58.447 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:58.708 [2024-11-26 19:26:11.112047] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:29:58.708 [2024-11-26 19:26:11.112106] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4051062 ] 00:29:58.708 I/O size of 131072 is greater than zero copy threshold (65536). 00:29:58.709 Zero copy mechanism will not be used. 00:29:58.709 [2024-11-26 19:26:11.201057] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:58.709 [2024-11-26 19:26:11.230762] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:59.281 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:59.281 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:29:59.281 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:59.281 19:26:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:29:59.543 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:29:59.543 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.543 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:29:59.543 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.543 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:59.543 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:29:59.804 nvme0n1 00:29:59.804 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:29:59.804 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.804 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:30:00.066 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.066 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:30:00.066 19:26:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:30:00.066 I/O size of 131072 is greater than zero copy threshold (65536). 00:30:00.066 Zero copy mechanism will not be used. 00:30:00.066 Running I/O for 2 seconds... 00:30:00.066 [2024-11-26 19:26:12.524240] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.524367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.524394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.531197] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.531331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.531350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.538230] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.538394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.538410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.544830] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.544991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.545008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.551044] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.551155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.551171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.557764] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.557852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.557873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.564155] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.564243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.564259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.569676] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.569813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.569829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.576168] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.576427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.576445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.582664] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.582784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.582800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.588699] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.588820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.588835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.594522] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.594630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.594646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.600613] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.600725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.600740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.606032] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.066 [2024-11-26 19:26:12.606197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.066 [2024-11-26 19:26:12.606212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.066 [2024-11-26 19:26:12.611367] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.611473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.611488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.616963] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.617066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.617086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.622419] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.622505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.622521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.627240] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.627300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.627316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.631808] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.631875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.631890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.636366] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.636449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.636465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.640784] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.640886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.640901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.645004] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.645088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.645103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.649613] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.649691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.649706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.654117] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.654199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.654214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.658338] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.658413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.658429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.662499] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.662567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.662582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.666450] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.666507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.666523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.670257] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.670317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.670331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.674038] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.674101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.674117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.677763] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.677821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.677836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.681507] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.681570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.681585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.067 [2024-11-26 19:26:12.685715] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.067 [2024-11-26 19:26:12.685781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.067 [2024-11-26 19:26:12.685796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.690830] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.690897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.690913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.696594] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.696655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.696669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.702413] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.702516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.702532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.708277] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.708341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.708357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.713698] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.713795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.713810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.718558] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.718623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.718638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.723040] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.723120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.723135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.727419] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.727490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.727505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.731765] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.731832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.731847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.736162] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.736245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.736263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.741052] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.741137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.741152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.745560] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.745625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.745640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.750075] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.750159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.750175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.754558] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.754629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.754644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.758932] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.759002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.759016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.762924] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.762997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.330 [2024-11-26 19:26:12.763011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.330 [2024-11-26 19:26:12.766828] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.330 [2024-11-26 19:26:12.766903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.766918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.772196] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.772393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.772409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.777446] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.777523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.777538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.781972] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.782072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.782086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.787820] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.788003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.788019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.795970] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.796148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.796164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.803332] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.803493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.803507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.811848] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.812014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.812030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.819993] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.820161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.820176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.827369] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.827524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.827539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.834685] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.834872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.834887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.842036] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.842190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.842204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.849315] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.849496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.849512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.855857] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.856046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.856061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.865286] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.865441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.865456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.873659] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.873804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.873819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.881244] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.881377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.881392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.887429] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.887544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.887559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.892834] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.893006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.893022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.898688] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.898868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.898887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.904215] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.904321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.904336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.909659] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.331 [2024-11-26 19:26:12.909769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.331 [2024-11-26 19:26:12.909784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.331 [2024-11-26 19:26:12.914809] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.332 [2024-11-26 19:26:12.914917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.332 [2024-11-26 19:26:12.914932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.332 [2024-11-26 19:26:12.920456] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.332 [2024-11-26 19:26:12.920567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.332 [2024-11-26 19:26:12.920581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.332 [2024-11-26 19:26:12.925855] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.332 [2024-11-26 19:26:12.925934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.332 [2024-11-26 19:26:12.925949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.332 [2024-11-26 19:26:12.931248] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.332 [2024-11-26 19:26:12.931384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.332 [2024-11-26 19:26:12.931398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.332 [2024-11-26 19:26:12.938475] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.332 [2024-11-26 19:26:12.938632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.332 [2024-11-26 19:26:12.938648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.332 [2024-11-26 19:26:12.945661] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.332 [2024-11-26 19:26:12.945807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.332 [2024-11-26 19:26:12.945822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.952154] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.952287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.952302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.957674] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.957733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.957748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.962656] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.962767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.962782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.971169] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.971446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.971462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.977378] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.977507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.977522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.983309] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.983465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.983481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.989324] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.989442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.989457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:12.995405] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:12.995569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:12.995585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:13.002850] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:13.003041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:13.003056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:13.012235] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.595 [2024-11-26 19:26:13.012388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.595 [2024-11-26 19:26:13.012402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.595 [2024-11-26 19:26:13.018441] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.018526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.018541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.024446] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.024555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.024570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.031372] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.031452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.031466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.037386] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.037502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.037517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.045296] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.045485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.045500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.052124] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.052282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.052297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.059609] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.059772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.059788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.066973] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.067081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.067100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.072852] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.072977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.072992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.078651] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.078761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.078776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.083877] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.083978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.083993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.089354] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.089615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.089630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.097435] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.097555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.097570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.103662] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.103790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.103805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.109578] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.109703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.109718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.115632] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.115803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.115817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.123660] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.123761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.123776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.129422] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.129601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.129617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.135215] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.135325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.135339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.140777] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.140900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.140915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.146180] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.146308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.596 [2024-11-26 19:26:13.146323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.596 [2024-11-26 19:26:13.152015] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.596 [2024-11-26 19:26:13.152145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.152160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.157882] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.157998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.158013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.163620] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.163738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.163753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.170177] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.170268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.170283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.176800] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.176878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.176893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.181777] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.181833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.181848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.186677] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.186752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.186767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.191328] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.191395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.191409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.195930] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.195997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.196012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.200437] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.200510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.200525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.205185] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.205257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.205271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.209771] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.209853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.209873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.597 [2024-11-26 19:26:13.214292] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.597 [2024-11-26 19:26:13.214363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.597 [2024-11-26 19:26:13.214381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.218930] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.219005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.219020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.223289] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.223356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.223371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.227539] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.227606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.227621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.232801] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.232857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.232877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.237592] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.237653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.237668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.241633] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.241713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.241728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.245747] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.245823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.245837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.249846] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.249923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.249938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.253926] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.254010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.254025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.258136] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.258218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.258233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.262358] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.262428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.262443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.266198] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.266260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.266274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.270324] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.270394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.270409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.274157] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.274222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.274237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.277906] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.277975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.277990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.281603] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.281661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.281675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.285283] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.285341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.285356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.288957] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.289021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.289036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.292690] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.292749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.292765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.296358] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.296431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.296446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.300036] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.300095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.300109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.303712] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.303780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.303795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.307384] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.307441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.307456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.311074] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.311144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.311159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.314765] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.314825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.314840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.318548] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.860 [2024-11-26 19:26:13.318613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.860 [2024-11-26 19:26:13.318630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.860 [2024-11-26 19:26:13.323855] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.324172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.324188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.327819] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.327893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.327908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.331570] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.331637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.331651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.335430] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.335546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.335561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.339905] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.339978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.339993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.344062] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.344155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.344170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.350268] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.350422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.350436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.357386] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.357497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.357512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.362449] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.362553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.362568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.368479] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.368546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.368561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.373295] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.373412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.373427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.378410] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.378490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.378505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.384342] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.384396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.384411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.389404] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.389494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.389509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.394248] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.394312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.394327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.399056] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.399171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.399186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.403724] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.403837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.403852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.408793] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.408872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.408887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.413984] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.414044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.414059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.418647] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.418712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.418726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.423492] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.423590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.423605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.428249] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.428309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.428324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.434312] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.434564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.434579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.439116] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.439203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.439219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.443473] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.443546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.443561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.447822] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.447887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.447904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.452170] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.861 [2024-11-26 19:26:13.452242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.861 [2024-11-26 19:26:13.452256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.861 [2024-11-26 19:26:13.456398] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.862 [2024-11-26 19:26:13.456481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.862 [2024-11-26 19:26:13.456496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.862 [2024-11-26 19:26:13.460427] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.862 [2024-11-26 19:26:13.460492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.862 [2024-11-26 19:26:13.460507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.862 [2024-11-26 19:26:13.464513] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.862 [2024-11-26 19:26:13.464589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.862 [2024-11-26 19:26:13.464604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:00.862 [2024-11-26 19:26:13.468644] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.862 [2024-11-26 19:26:13.468710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.862 [2024-11-26 19:26:13.468725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:00.862 [2024-11-26 19:26:13.472498] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.862 [2024-11-26 19:26:13.472570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.862 [2024-11-26 19:26:13.472585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:00.862 [2024-11-26 19:26:13.476293] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.862 [2024-11-26 19:26:13.476360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.862 [2024-11-26 19:26:13.476374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:00.862 [2024-11-26 19:26:13.480060] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:00.862 [2024-11-26 19:26:13.480123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:00.862 [2024-11-26 19:26:13.480137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.125 [2024-11-26 19:26:13.483821] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.483901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.483916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.487557] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.487618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.487632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.491331] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.491390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.491404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.495016] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.495076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.495091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.498670] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.498726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.498741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.502332] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.502387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.502402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.506009] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.506063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.506078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.509678] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.509740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.509755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.513351] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.513408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.513423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.517087] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.517137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.517151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.126 5861.00 IOPS, 732.62 MiB/s [2024-11-26T18:26:13.751Z] [2024-11-26 19:26:13.521899] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.521972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.521987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.525891] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.525981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.525996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.529886] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.529944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.529958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.533656] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.533736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.533750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.538454] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.538673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.538690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.547377] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.547595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.547610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.552561] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.552752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.552768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.558724] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.558909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.558931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.563258] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.563449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.563465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.568620] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.568807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.568822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.573753] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.573947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.573963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.579941] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.580167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.580182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.589714] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.589997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.590013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.597872] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.598079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.598095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.604233] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.604457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.604473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.610184] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.610377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.610393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.126 [2024-11-26 19:26:13.615413] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.126 [2024-11-26 19:26:13.615603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.126 [2024-11-26 19:26:13.615618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.620289] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.620500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.620515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.624824] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.625136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.625152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.629795] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.629991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.630007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.634369] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.634666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.634682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.639537] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.639725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.639741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.643886] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.644070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.644087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.649216] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.649398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.649413] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.653392] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.653577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.653593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.658116] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.658428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.658443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.662631] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.662813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.662829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.666971] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.667160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.667175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.671211] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.671395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.671411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.676009] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.676224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.676240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.680078] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.680262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.680278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.683937] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.684118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.684134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.687580] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.687787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.687802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.691338] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.691529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.691548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.695106] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.695284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.695300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.698972] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.699156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.699172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.702564] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.702752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.702768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.706133] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.706315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.706331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.709726] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.709989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.710004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.713406] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.713588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.713604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.716941] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.717130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.717145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.720480] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.720667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.720682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.724126] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.724317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.724333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.727663] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.127 [2024-11-26 19:26:13.727841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.127 [2024-11-26 19:26:13.727858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.127 [2024-11-26 19:26:13.731243] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.128 [2024-11-26 19:26:13.731438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.128 [2024-11-26 19:26:13.731453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.128 [2024-11-26 19:26:13.734770] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.128 [2024-11-26 19:26:13.734956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.128 [2024-11-26 19:26:13.734972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.128 [2024-11-26 19:26:13.738271] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.128 [2024-11-26 19:26:13.738455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.128 [2024-11-26 19:26:13.738471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.128 [2024-11-26 19:26:13.741801] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.128 [2024-11-26 19:26:13.741998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.128 [2024-11-26 19:26:13.742015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.128 [2024-11-26 19:26:13.745301] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.128 [2024-11-26 19:26:13.745487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.128 [2024-11-26 19:26:13.745502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.391 [2024-11-26 19:26:13.748802] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.748999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.749014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.752312] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.752499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.752515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.755807] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.755999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.756015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.759404] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.759593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.759609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.762997] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.763136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.763151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.766708] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.766878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.766894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.772359] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.772542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.772558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.777145] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.777314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.777330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.781338] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.781503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.781519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.786053] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.786213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.786228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.790368] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.790603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.790622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.794731] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.795037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.795053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.799441] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.799709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.799725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.803750] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.803942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.803958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.807580] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.807743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.807759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.811486] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.811651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.811667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.815273] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.815437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.815453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.818972] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.819153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.819168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.822467] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.822629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.822646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.825966] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.826282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.826297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.829888] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.830060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.830075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.833662] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.833949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.833965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.837523] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.837695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.837711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.841032] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.841204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.841220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.844484] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.844666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.844681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.847957] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.848132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.392 [2024-11-26 19:26:13.848147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.392 [2024-11-26 19:26:13.851394] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.392 [2024-11-26 19:26:13.851562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.851578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.854825] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.855002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.855017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.858274] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.858440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.858455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.861733] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.861906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.861922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.865185] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.865356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.865372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.868664] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.868845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.868865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.872138] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.872307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.872323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.875583] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.875751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.875767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.879034] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.879195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.879210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.882516] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.882683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.882699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.886153] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.886323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.886342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.889601] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.889762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.889778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.893069] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.893235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.893251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.896534] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.896700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.896716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.899976] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.900146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.900161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.903446] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.903603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.903619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.906908] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.907074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.907091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.910362] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.910528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.910544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.913822] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.913992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.914008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.917294] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.917466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.917482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.920764] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.920935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.920951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.924229] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.924401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.924417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.928120] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.928287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.928303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.931588] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.931757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.931773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.935048] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.935210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.935225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.938511] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.938678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.938694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.941986] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.942155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.393 [2024-11-26 19:26:13.942170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.393 [2024-11-26 19:26:13.945430] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.393 [2024-11-26 19:26:13.945602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.945618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.949009] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.949175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.949191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.952769] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.952948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.952964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.957122] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.957292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.957308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.961261] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.961435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.961450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.965158] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.965327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.965343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.969541] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.969723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.969738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.974292] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.974572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.974588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.979749] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.979922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.979938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.984601] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.984922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:32 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.984941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.990385] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.990601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.990617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:13.996512] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:13.996815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:13.996830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:14.001951] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:14.002119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:14.002135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.394 [2024-11-26 19:26:14.007078] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.394 [2024-11-26 19:26:14.007249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.394 [2024-11-26 19:26:14.007266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.656 [2024-11-26 19:26:14.012268] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.656 [2024-11-26 19:26:14.012440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.656 [2024-11-26 19:26:14.012456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.656 [2024-11-26 19:26:14.018165] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.656 [2024-11-26 19:26:14.018351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.656 [2024-11-26 19:26:14.018366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.656 [2024-11-26 19:26:14.024737] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.656 [2024-11-26 19:26:14.024831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.656 [2024-11-26 19:26:14.024846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.032294] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.032795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.032811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.042159] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.042404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.042420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.052734] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.053082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.053098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.062187] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.062659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.062675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.072674] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.072965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.072980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.080207] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.080412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.080428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.087737] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.088114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.088130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.098180] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.098428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.098443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.107205] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.107429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.107445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.115247] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.115463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.115479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.123256] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.123534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.123550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.130152] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.130393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.130409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.136575] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.136899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.136915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.145615] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.145875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.145891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.153734] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.154152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.154168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.159242] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.159415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.159431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.165044] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.165220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.165235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.170541] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.170873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.170889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.176287] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.176460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.176479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.180579] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.180868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.180883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.185509] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.657 [2024-11-26 19:26:14.185818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.657 [2024-11-26 19:26:14.185833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.657 [2024-11-26 19:26:14.190161] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.190441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.190457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.194494] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.194664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.194680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.198695] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.199016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.199032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.202786] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.202963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.202979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.206900] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.207071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.207087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.210677] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.210839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.210855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.214313] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.214473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.214489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.218038] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.218207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.218222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.221955] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.222125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.222141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.225430] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.225601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.225616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.229062] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.229231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.229246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.232926] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.233089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.233104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.236401] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.236571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.236586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.239879] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.240046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.240062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.243373] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.243660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.243676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.247071] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.247247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.247263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.250928] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.251090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.251106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.257554] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.257827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.257842] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.264351] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.264517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.264533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.268267] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.268473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.268489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.272243] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.272395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.658 [2024-11-26 19:26:14.272410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.658 [2024-11-26 19:26:14.276383] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.658 [2024-11-26 19:26:14.276538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.659 [2024-11-26 19:26:14.276553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.283694] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.283927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.283943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.289623] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.289775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.289793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.294553] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.294746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.294761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.300529] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.300699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.300715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.307988] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.308265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.308281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.315204] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.315376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.315391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.320756] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.320928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.320944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.326133] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.326293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.326308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.330970] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.331134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.331149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.336221] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.336496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.336512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.344333] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.344466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.344481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.349495] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.349692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.349708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.354976] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.355139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.355153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.360533] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.360722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.360738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.365699] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.365890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.365905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.370566] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.370721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.370737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.375951] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.376115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.376131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.380503] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.380657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.380673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.385014] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.385172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.385188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.389819] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.390016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.390032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.398617] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.398880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.398896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.404843] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.405008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.405023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.410851] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.411058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.921 [2024-11-26 19:26:14.411074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.921 [2024-11-26 19:26:14.416260] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.921 [2024-11-26 19:26:14.416423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.416438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.421223] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.421385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.421401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.426177] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.426336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.426352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.431012] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.431161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.431177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.434963] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.435129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.435147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.438735] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.438892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.438907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.442574] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.442723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.442738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.446725] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.447036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.447052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.450749] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.451003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.451019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.454761] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.455048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.455064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.458486] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.458660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.458676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.461805] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.461979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.461995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.465136] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.465274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.465290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.468429] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.468610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.468626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.471748] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.471909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.471924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.475059] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.475218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.475234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.478366] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.478539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.478554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.481697] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.481872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.481888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.484980] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.485268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.485283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.488435] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.488566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.488580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.491721] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.491869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.491885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.495029] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.495174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.495189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.498310] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.498462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.498477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.501593] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.501726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.501741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.505078] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.505233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.505247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.508352] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.508512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.508528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.511674] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.511830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.922 [2024-11-26 19:26:14.511845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:30:01.922 [2024-11-26 19:26:14.514999] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.922 [2024-11-26 19:26:14.515156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.923 [2024-11-26 19:26:14.515171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:30:01.923 [2024-11-26 19:26:14.518310] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1d9ed60) with pdu=0x200016eff3c8 00:30:01.923 [2024-11-26 19:26:14.518464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:01.923 [2024-11-26 19:26:14.518479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:30:01.923 6253.50 IOPS, 781.69 MiB/s 00:30:01.923 Latency(us) 00:30:01.923 [2024-11-26T18:26:14.548Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:01.923 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:30:01.923 nvme0n1 : 2.00 6252.31 781.54 0.00 0.00 2555.64 1570.13 12615.68 00:30:01.923 [2024-11-26T18:26:14.548Z] =================================================================================================================== 00:30:01.923 [2024-11-26T18:26:14.548Z] Total : 6252.31 781.54 0.00 0.00 2555.64 1570.13 12615.68 00:30:01.923 { 00:30:01.923 "results": [ 00:30:01.923 { 00:30:01.923 "job": "nvme0n1", 00:30:01.923 "core_mask": "0x2", 00:30:01.923 "workload": "randwrite", 00:30:01.923 "status": "finished", 00:30:01.923 "queue_depth": 16, 00:30:01.923 "io_size": 131072, 00:30:01.923 "runtime": 2.002941, 00:30:01.923 "iops": 6252.305984050454, 00:30:01.923 "mibps": 781.5382480063067, 00:30:01.923 "io_failed": 0, 00:30:01.923 "io_timeout": 0, 00:30:01.923 "avg_latency_us": 2555.637594825521, 00:30:01.923 "min_latency_us": 1570.1333333333334, 00:30:01.923 "max_latency_us": 12615.68 00:30:01.923 } 00:30:01.923 ], 00:30:01.923 "core_count": 1 00:30:01.923 } 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:30:02.183 | .driver_specific 00:30:02.183 | .nvme_error 00:30:02.183 | .status_code 00:30:02.183 | .command_transient_transport_error' 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 404 > 0 )) 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 4051062 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 4051062 ']' 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 4051062 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4051062 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4051062' 00:30:02.183 killing process with pid 4051062 00:30:02.183 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 4051062 00:30:02.183 Received shutdown signal, test time was about 2.000000 seconds 00:30:02.183 00:30:02.184 Latency(us) 00:30:02.184 [2024-11-26T18:26:14.809Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:02.184 [2024-11-26T18:26:14.809Z] =================================================================================================================== 00:30:02.184 [2024-11-26T18:26:14.809Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:30:02.184 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 4051062 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@116 -- # killprocess 4048865 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 4048865 ']' 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 4048865 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4048865 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4048865' 00:30:02.446 killing process with pid 4048865 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 4048865 00:30:02.446 19:26:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 4048865 00:30:02.446 00:30:02.446 real 0m15.648s 00:30:02.446 user 0m30.726s 00:30:02.446 sys 0m3.476s 00:30:02.446 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:02.446 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:30:02.446 ************************************ 00:30:02.446 END TEST nvmf_digest_error 00:30:02.446 ************************************ 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@150 -- # nvmftestfini 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@516 -- # nvmfcleanup 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@121 -- # sync 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@124 -- # set +e 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:02.707 rmmod nvme_tcp 00:30:02.707 rmmod nvme_fabrics 00:30:02.707 rmmod nvme_keyring 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@128 -- # set -e 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@129 -- # return 0 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@517 -- # '[' -n 4048865 ']' 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@518 -- # killprocess 4048865 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@954 -- # '[' -z 4048865 ']' 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@958 -- # kill -0 4048865 00:30:02.707 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (4048865) - No such process 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@981 -- # echo 'Process with pid 4048865 is not found' 00:30:02.707 Process with pid 4048865 is not found 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@297 -- # iptr 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@791 -- # iptables-save 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@791 -- # iptables-restore 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:02.707 19:26:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:04.621 19:26:17 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:04.883 00:30:04.883 real 0m42.976s 00:30:04.883 user 1m5.659s 00:30:04.883 sys 0m13.463s 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:30:04.883 ************************************ 00:30:04.883 END TEST nvmf_digest 00:30:04.883 ************************************ 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@36 -- # [[ 0 -eq 1 ]] 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@41 -- # [[ 0 -eq 1 ]] 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@46 -- # [[ phy == phy ]] 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@47 -- # run_test nvmf_bdevperf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:30:04.883 ************************************ 00:30:04.883 START TEST nvmf_bdevperf 00:30:04.883 ************************************ 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:30:04.883 * Looking for test storage... 00:30:04.883 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1693 -- # lcov --version 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # IFS=.-: 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # read -ra ver1 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # IFS=.-: 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # read -ra ver2 00:30:04.883 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@338 -- # local 'op=<' 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@340 -- # ver1_l=2 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@341 -- # ver2_l=1 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@344 -- # case "$op" in 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@345 -- # : 1 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # decimal 1 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=1 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 1 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # ver1[v]=1 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # decimal 2 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=2 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 2 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # ver2[v]=2 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # return 0 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:30:05.145 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:05.145 --rc genhtml_branch_coverage=1 00:30:05.145 --rc genhtml_function_coverage=1 00:30:05.145 --rc genhtml_legend=1 00:30:05.145 --rc geninfo_all_blocks=1 00:30:05.145 --rc geninfo_unexecuted_blocks=1 00:30:05.145 00:30:05.145 ' 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:30:05.145 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:05.145 --rc genhtml_branch_coverage=1 00:30:05.145 --rc genhtml_function_coverage=1 00:30:05.145 --rc genhtml_legend=1 00:30:05.145 --rc geninfo_all_blocks=1 00:30:05.145 --rc geninfo_unexecuted_blocks=1 00:30:05.145 00:30:05.145 ' 00:30:05.145 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:30:05.145 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:05.146 --rc genhtml_branch_coverage=1 00:30:05.146 --rc genhtml_function_coverage=1 00:30:05.146 --rc genhtml_legend=1 00:30:05.146 --rc geninfo_all_blocks=1 00:30:05.146 --rc geninfo_unexecuted_blocks=1 00:30:05.146 00:30:05.146 ' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:30:05.146 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:05.146 --rc genhtml_branch_coverage=1 00:30:05.146 --rc genhtml_function_coverage=1 00:30:05.146 --rc genhtml_legend=1 00:30:05.146 --rc geninfo_all_blocks=1 00:30:05.146 --rc geninfo_unexecuted_blocks=1 00:30:05.146 00:30:05.146 ' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # uname -s 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@15 -- # shopt -s extglob 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@5 -- # export PATH 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@51 -- # : 0 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:30:05.146 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@11 -- # MALLOC_BDEV_SIZE=64 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@24 -- # nvmftestinit 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@476 -- # prepare_net_devs 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@438 -- # local -g is_hw=no 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # remove_spdk_ns 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@309 -- # xtrace_disable 00:30:05.146 19:26:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # pci_devs=() 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # net_devs=() 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # e810=() 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # local -ga e810 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # x722=() 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # local -ga x722 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # mlx=() 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # local -ga mlx 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:30:13.328 Found 0000:31:00.0 (0x8086 - 0x159b) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:30:13.328 Found 0000:31:00.1 (0x8086 - 0x159b) 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:13.328 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:30:13.329 Found net devices under 0000:31:00.0: cvl_0_0 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:30:13.329 Found net devices under 0000:31:00.1: cvl_0_1 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # is_hw=yes 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:13.329 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:13.329 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.649 ms 00:30:13.329 00:30:13.329 --- 10.0.0.2 ping statistics --- 00:30:13.329 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:13.329 rtt min/avg/max/mdev = 0.649/0.649/0.649/0.000 ms 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:13.329 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:13.329 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.289 ms 00:30:13.329 00:30:13.329 --- 10.0.0.1 ping statistics --- 00:30:13.329 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:13.329 rtt min/avg/max/mdev = 0.289/0.289/0.289/0.000 ms 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@450 -- # return 0 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:30:13.329 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@25 -- # tgt_init 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@726 -- # xtrace_disable 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # nvmfpid=4056635 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # waitforlisten 4056635 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # '[' -z 4056635 ']' 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:13.589 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:13.589 19:26:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.589 [2024-11-26 19:26:26.025335] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:30:13.589 [2024-11-26 19:26:26.025386] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:13.589 [2024-11-26 19:26:26.121082] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:30:13.589 [2024-11-26 19:26:26.150827] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:13.589 [2024-11-26 19:26:26.150858] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:13.589 [2024-11-26 19:26:26.150869] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:13.589 [2024-11-26 19:26:26.150874] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:13.589 [2024-11-26 19:26:26.150878] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:13.589 [2024-11-26 19:26:26.152028] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:30:13.589 [2024-11-26 19:26:26.152394] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:30:13.589 [2024-11-26 19:26:26.152394] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@868 -- # return 0 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@732 -- # xtrace_disable 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.849 [2024-11-26 19:26:26.271912] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.849 Malloc0 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.849 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:13.850 [2024-11-26 19:26:26.336773] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 128 -o 4096 -w verify -t 1 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # gen_nvmf_target_json 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # config=() 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # local subsystem config 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:30:13.850 { 00:30:13.850 "params": { 00:30:13.850 "name": "Nvme$subsystem", 00:30:13.850 "trtype": "$TEST_TRANSPORT", 00:30:13.850 "traddr": "$NVMF_FIRST_TARGET_IP", 00:30:13.850 "adrfam": "ipv4", 00:30:13.850 "trsvcid": "$NVMF_PORT", 00:30:13.850 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:30:13.850 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:30:13.850 "hdgst": ${hdgst:-false}, 00:30:13.850 "ddgst": ${ddgst:-false} 00:30:13.850 }, 00:30:13.850 "method": "bdev_nvme_attach_controller" 00:30:13.850 } 00:30:13.850 EOF 00:30:13.850 )") 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # cat 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # jq . 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@585 -- # IFS=, 00:30:13.850 19:26:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:30:13.850 "params": { 00:30:13.850 "name": "Nvme1", 00:30:13.850 "trtype": "tcp", 00:30:13.850 "traddr": "10.0.0.2", 00:30:13.850 "adrfam": "ipv4", 00:30:13.850 "trsvcid": "4420", 00:30:13.850 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:30:13.850 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:30:13.850 "hdgst": false, 00:30:13.850 "ddgst": false 00:30:13.850 }, 00:30:13.850 "method": "bdev_nvme_attach_controller" 00:30:13.850 }' 00:30:13.850 [2024-11-26 19:26:26.401188] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:30:13.850 [2024-11-26 19:26:26.401239] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4056667 ] 00:30:14.109 [2024-11-26 19:26:26.478740] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:14.109 [2024-11-26 19:26:26.514832] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:14.369 Running I/O for 1 seconds... 00:30:15.310 11146.00 IOPS, 43.54 MiB/s 00:30:15.310 Latency(us) 00:30:15.310 [2024-11-26T18:26:27.935Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:15.310 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:30:15.310 Verification LBA range: start 0x0 length 0x4000 00:30:15.310 Nvme1n1 : 1.01 11225.52 43.85 0.00 0.00 11342.32 2389.33 10813.44 00:30:15.310 [2024-11-26T18:26:27.935Z] =================================================================================================================== 00:30:15.310 [2024-11-26T18:26:27.935Z] Total : 11225.52 43.85 0.00 0.00 11342.32 2389.33 10813.44 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@30 -- # bdevperfpid=4057001 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@32 -- # sleep 3 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -q 128 -o 4096 -w verify -t 15 -f 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # gen_nvmf_target_json 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # config=() 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # local subsystem config 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:30:15.310 { 00:30:15.310 "params": { 00:30:15.310 "name": "Nvme$subsystem", 00:30:15.310 "trtype": "$TEST_TRANSPORT", 00:30:15.310 "traddr": "$NVMF_FIRST_TARGET_IP", 00:30:15.310 "adrfam": "ipv4", 00:30:15.310 "trsvcid": "$NVMF_PORT", 00:30:15.310 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:30:15.310 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:30:15.310 "hdgst": ${hdgst:-false}, 00:30:15.310 "ddgst": ${ddgst:-false} 00:30:15.310 }, 00:30:15.310 "method": "bdev_nvme_attach_controller" 00:30:15.310 } 00:30:15.310 EOF 00:30:15.310 )") 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # cat 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # jq . 00:30:15.310 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@585 -- # IFS=, 00:30:15.311 19:26:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:30:15.311 "params": { 00:30:15.311 "name": "Nvme1", 00:30:15.311 "trtype": "tcp", 00:30:15.311 "traddr": "10.0.0.2", 00:30:15.311 "adrfam": "ipv4", 00:30:15.311 "trsvcid": "4420", 00:30:15.311 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:30:15.311 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:30:15.311 "hdgst": false, 00:30:15.311 "ddgst": false 00:30:15.311 }, 00:30:15.311 "method": "bdev_nvme_attach_controller" 00:30:15.311 }' 00:30:15.571 [2024-11-26 19:26:27.966122] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:30:15.571 [2024-11-26 19:26:27.966180] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4057001 ] 00:30:15.571 [2024-11-26 19:26:28.042307] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:15.571 [2024-11-26 19:26:28.077357] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:15.831 Running I/O for 15 seconds... 00:30:17.717 11241.00 IOPS, 43.91 MiB/s [2024-11-26T18:26:31.289Z] 11170.50 IOPS, 43.63 MiB/s [2024-11-26T18:26:31.289Z] 19:26:30 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@33 -- # kill -9 4056635 00:30:18.664 19:26:30 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@35 -- # sleep 3 00:30:18.664 [2024-11-26 19:26:30.932081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:109544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.664 [2024-11-26 19:26:30.932122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:109608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:109616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:109624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:109632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:109640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:109648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:109656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:109664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:109672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:109680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:109688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:109696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:109704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:109712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:109720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:109728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:109736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:109744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:109752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.664 [2024-11-26 19:26:30.932517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.664 [2024-11-26 19:26:30.932526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:109760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:109768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:109776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:109784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:109792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:109800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:109808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:109816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:109824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:109832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:109840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:109848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:109856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:109864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:109872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:109880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:109888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:109896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:109904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:109912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:109920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:109928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:109936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:109944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:109952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:109960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:109968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.932987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.932997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:109976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.933004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.933014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:109984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.933021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.933031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:109992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.665 [2024-11-26 19:26:30.933038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.665 [2024-11-26 19:26:30.933048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:110000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:110008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:110016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:110024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:110032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:110040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:110048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:110056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:110064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:110072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:110080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:110088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:110096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:110104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:110112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933296] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:110120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:110128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:109552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.666 [2024-11-26 19:26:30.933349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:109560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.666 [2024-11-26 19:26:30.933368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:109568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.666 [2024-11-26 19:26:30.933386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:109576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.666 [2024-11-26 19:26:30.933403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:109584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.666 [2024-11-26 19:26:30.933421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:109592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.666 [2024-11-26 19:26:30.933438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:109600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:18.666 [2024-11-26 19:26:30.933456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:110136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:110144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:110152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:110160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:110168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:110176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:110184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.666 [2024-11-26 19:26:30.933588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:110192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.666 [2024-11-26 19:26:30.933596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:110200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:110208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:110216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:110224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:110232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:110240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933707] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:110248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:110256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:110264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:110272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:110280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:110288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:110296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:110304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:110312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:110320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:110328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:110336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:110344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:110352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:110360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:110368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.933988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:110376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.933995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:110384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:110392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:110400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:110408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:110416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:110424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:110432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:110440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.667 [2024-11-26 19:26:30.934133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.667 [2024-11-26 19:26:30.934143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:110448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:110456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:110464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:110472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:110480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934228] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:110488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:110496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:110504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:110512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:110520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:110528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:110536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:110544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:110552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:30:18.668 [2024-11-26 19:26:30.934372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.934380] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d80970 is same with the state(6) to be set 00:30:18.668 [2024-11-26 19:26:30.934389] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:30:18.668 [2024-11-26 19:26:30.934395] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:30:18.668 [2024-11-26 19:26:30.934402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:110560 len:8 PRP1 0x0 PRP2 0x0 00:30:18.668 [2024-11-26 19:26:30.934410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:18.668 [2024-11-26 19:26:30.937951] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.668 [2024-11-26 19:26:30.938004] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.668 [2024-11-26 19:26:30.938813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.668 [2024-11-26 19:26:30.938830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.668 [2024-11-26 19:26:30.938843] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.668 [2024-11-26 19:26:30.939071] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.668 [2024-11-26 19:26:30.939293] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.668 [2024-11-26 19:26:30.939302] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.668 [2024-11-26 19:26:30.939311] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.668 [2024-11-26 19:26:30.939320] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.668 [2024-11-26 19:26:30.952163] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.668 [2024-11-26 19:26:30.952829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.668 [2024-11-26 19:26:30.952878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.668 [2024-11-26 19:26:30.952890] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.668 [2024-11-26 19:26:30.953132] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.668 [2024-11-26 19:26:30.953358] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.668 [2024-11-26 19:26:30.953366] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.668 [2024-11-26 19:26:30.953376] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.668 [2024-11-26 19:26:30.953384] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.668 [2024-11-26 19:26:30.966017] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.668 [2024-11-26 19:26:30.966609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.668 [2024-11-26 19:26:30.966629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.668 [2024-11-26 19:26:30.966637] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.668 [2024-11-26 19:26:30.966858] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.668 [2024-11-26 19:26:30.967085] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.668 [2024-11-26 19:26:30.967092] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.668 [2024-11-26 19:26:30.967100] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.668 [2024-11-26 19:26:30.967107] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.668 [2024-11-26 19:26:30.979951] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.668 [2024-11-26 19:26:30.980610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.668 [2024-11-26 19:26:30.980648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.668 [2024-11-26 19:26:30.980659] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.668 [2024-11-26 19:26:30.980908] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.668 [2024-11-26 19:26:30.981141] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.668 [2024-11-26 19:26:30.981150] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.668 [2024-11-26 19:26:30.981159] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.668 [2024-11-26 19:26:30.981167] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.668 [2024-11-26 19:26:30.993808] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.668 [2024-11-26 19:26:30.994475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.668 [2024-11-26 19:26:30.994513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.668 [2024-11-26 19:26:30.994524] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.669 [2024-11-26 19:26:30.994764] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.669 [2024-11-26 19:26:30.994998] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.669 [2024-11-26 19:26:30.995007] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.669 [2024-11-26 19:26:30.995015] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.669 [2024-11-26 19:26:30.995024] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.669 [2024-11-26 19:26:31.007656] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.669 [2024-11-26 19:26:31.008215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.669 [2024-11-26 19:26:31.008234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.669 [2024-11-26 19:26:31.008243] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.669 [2024-11-26 19:26:31.008464] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.669 [2024-11-26 19:26:31.008684] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.669 [2024-11-26 19:26:31.008692] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.669 [2024-11-26 19:26:31.008699] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.669 [2024-11-26 19:26:31.008706] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.669 [2024-11-26 19:26:31.021543] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.669 [2024-11-26 19:26:31.022188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.669 [2024-11-26 19:26:31.022226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.669 [2024-11-26 19:26:31.022237] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.669 [2024-11-26 19:26:31.022477] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.669 [2024-11-26 19:26:31.022702] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.669 [2024-11-26 19:26:31.022711] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.669 [2024-11-26 19:26:31.022723] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.669 [2024-11-26 19:26:31.022732] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.669 [2024-11-26 19:26:31.035365] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.669 [2024-11-26 19:26:31.035977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.669 [2024-11-26 19:26:31.036015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.669 [2024-11-26 19:26:31.036027] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.669 [2024-11-26 19:26:31.036269] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.669 [2024-11-26 19:26:31.036493] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.669 [2024-11-26 19:26:31.036502] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.669 [2024-11-26 19:26:31.036509] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.669 [2024-11-26 19:26:31.036517] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.669 [2024-11-26 19:26:31.049363] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.669 [2024-11-26 19:26:31.049969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.669 [2024-11-26 19:26:31.050006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.669 [2024-11-26 19:26:31.050018] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.669 [2024-11-26 19:26:31.050260] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.669 [2024-11-26 19:26:31.050484] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.669 [2024-11-26 19:26:31.050492] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.669 [2024-11-26 19:26:31.050501] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.669 [2024-11-26 19:26:31.050508] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.669 [2024-11-26 19:26:31.063357] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.669 [2024-11-26 19:26:31.063957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.669 [2024-11-26 19:26:31.063996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.669 [2024-11-26 19:26:31.064007] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.669 [2024-11-26 19:26:31.064246] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.669 [2024-11-26 19:26:31.064471] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.669 [2024-11-26 19:26:31.064479] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.669 [2024-11-26 19:26:31.064487] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.669 [2024-11-26 19:26:31.064494] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.669 [2024-11-26 19:26:31.077361] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.669 [2024-11-26 19:26:31.077967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.669 [2024-11-26 19:26:31.078005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.669 [2024-11-26 19:26:31.078018] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.669 [2024-11-26 19:26:31.078261] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.669 [2024-11-26 19:26:31.078485] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.669 [2024-11-26 19:26:31.078494] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.078501] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.670 [2024-11-26 19:26:31.078509] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.670 [2024-11-26 19:26:31.091371] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.670 [2024-11-26 19:26:31.091969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.670 [2024-11-26 19:26:31.092008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.670 [2024-11-26 19:26:31.092020] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.670 [2024-11-26 19:26:31.092264] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.670 [2024-11-26 19:26:31.092488] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.670 [2024-11-26 19:26:31.092497] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.092504] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.670 [2024-11-26 19:26:31.092512] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.670 [2024-11-26 19:26:31.105356] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.670 [2024-11-26 19:26:31.106098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.670 [2024-11-26 19:26:31.106135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.670 [2024-11-26 19:26:31.106146] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.670 [2024-11-26 19:26:31.106387] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.670 [2024-11-26 19:26:31.106611] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.670 [2024-11-26 19:26:31.106620] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.106628] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.670 [2024-11-26 19:26:31.106635] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.670 [2024-11-26 19:26:31.119280] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.670 [2024-11-26 19:26:31.119962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.670 [2024-11-26 19:26:31.120000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.670 [2024-11-26 19:26:31.120017] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.670 [2024-11-26 19:26:31.120261] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.670 [2024-11-26 19:26:31.120485] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.670 [2024-11-26 19:26:31.120493] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.120502] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.670 [2024-11-26 19:26:31.120510] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.670 [2024-11-26 19:26:31.133144] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.670 [2024-11-26 19:26:31.133805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.670 [2024-11-26 19:26:31.133843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.670 [2024-11-26 19:26:31.133856] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.670 [2024-11-26 19:26:31.134108] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.670 [2024-11-26 19:26:31.134332] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.670 [2024-11-26 19:26:31.134341] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.134350] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.670 [2024-11-26 19:26:31.134357] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.670 [2024-11-26 19:26:31.146994] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.670 [2024-11-26 19:26:31.147658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.670 [2024-11-26 19:26:31.147696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.670 [2024-11-26 19:26:31.147707] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.670 [2024-11-26 19:26:31.147959] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.670 [2024-11-26 19:26:31.148184] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.670 [2024-11-26 19:26:31.148193] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.148201] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.670 [2024-11-26 19:26:31.148209] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.670 [2024-11-26 19:26:31.160837] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.670 [2024-11-26 19:26:31.161487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.670 [2024-11-26 19:26:31.161524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.670 [2024-11-26 19:26:31.161535] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.670 [2024-11-26 19:26:31.161776] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.670 [2024-11-26 19:26:31.162013] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.670 [2024-11-26 19:26:31.162023] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.162031] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.670 [2024-11-26 19:26:31.162039] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.670 [2024-11-26 19:26:31.174682] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.670 [2024-11-26 19:26:31.175343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.670 [2024-11-26 19:26:31.175381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.670 [2024-11-26 19:26:31.175391] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.670 [2024-11-26 19:26:31.175631] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.670 [2024-11-26 19:26:31.175856] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.670 [2024-11-26 19:26:31.175871] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.670 [2024-11-26 19:26:31.175880] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.175888] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.671 [2024-11-26 19:26:31.188525] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.671 [2024-11-26 19:26:31.189175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.671 [2024-11-26 19:26:31.189213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.671 [2024-11-26 19:26:31.189226] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.671 [2024-11-26 19:26:31.189467] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.671 [2024-11-26 19:26:31.189691] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.671 [2024-11-26 19:26:31.189702] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.671 [2024-11-26 19:26:31.189712] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.189721] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.671 [2024-11-26 19:26:31.202361] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.671 [2024-11-26 19:26:31.203049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.671 [2024-11-26 19:26:31.203087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.671 [2024-11-26 19:26:31.203098] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.671 [2024-11-26 19:26:31.203338] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.671 [2024-11-26 19:26:31.203563] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.671 [2024-11-26 19:26:31.203572] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.671 [2024-11-26 19:26:31.203585] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.203595] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.671 [2024-11-26 19:26:31.216232] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.671 [2024-11-26 19:26:31.216824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.671 [2024-11-26 19:26:31.216844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.671 [2024-11-26 19:26:31.216853] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.671 [2024-11-26 19:26:31.217079] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.671 [2024-11-26 19:26:31.217300] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.671 [2024-11-26 19:26:31.217309] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.671 [2024-11-26 19:26:31.217316] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.217323] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.671 [2024-11-26 19:26:31.230158] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.671 [2024-11-26 19:26:31.230736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.671 [2024-11-26 19:26:31.230753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.671 [2024-11-26 19:26:31.230760] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.671 [2024-11-26 19:26:31.230985] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.671 [2024-11-26 19:26:31.231206] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.671 [2024-11-26 19:26:31.231213] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.671 [2024-11-26 19:26:31.231220] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.231227] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.671 [2024-11-26 19:26:31.244053] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.671 [2024-11-26 19:26:31.244714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.671 [2024-11-26 19:26:31.244750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.671 [2024-11-26 19:26:31.244761] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.671 [2024-11-26 19:26:31.245011] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.671 [2024-11-26 19:26:31.245236] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.671 [2024-11-26 19:26:31.245245] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.671 [2024-11-26 19:26:31.245253] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.245261] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.671 [2024-11-26 19:26:31.257888] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.671 [2024-11-26 19:26:31.258556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.671 [2024-11-26 19:26:31.258593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.671 [2024-11-26 19:26:31.258604] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.671 [2024-11-26 19:26:31.258844] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.671 [2024-11-26 19:26:31.259078] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.671 [2024-11-26 19:26:31.259088] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.671 [2024-11-26 19:26:31.259096] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.259104] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.671 [2024-11-26 19:26:31.271742] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.671 [2024-11-26 19:26:31.272435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.671 [2024-11-26 19:26:31.272473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.671 [2024-11-26 19:26:31.272483] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.671 [2024-11-26 19:26:31.272723] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.671 [2024-11-26 19:26:31.272958] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.671 [2024-11-26 19:26:31.272968] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.671 [2024-11-26 19:26:31.272975] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.671 [2024-11-26 19:26:31.272983] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.934 10025.67 IOPS, 39.16 MiB/s [2024-11-26T18:26:31.559Z] [2024-11-26 19:26:31.285597] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.934 [2024-11-26 19:26:31.286273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.934 [2024-11-26 19:26:31.286311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.934 [2024-11-26 19:26:31.286322] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.934 [2024-11-26 19:26:31.286562] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.934 [2024-11-26 19:26:31.286787] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.934 [2024-11-26 19:26:31.286795] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.934 [2024-11-26 19:26:31.286803] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.934 [2024-11-26 19:26:31.286811] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.299451] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.300126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.300164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.935 [2024-11-26 19:26:31.300180] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.935 [2024-11-26 19:26:31.300420] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.935 [2024-11-26 19:26:31.300644] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.935 [2024-11-26 19:26:31.300653] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.935 [2024-11-26 19:26:31.300660] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.935 [2024-11-26 19:26:31.300668] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.313304] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.313977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.314015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.935 [2024-11-26 19:26:31.314027] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.935 [2024-11-26 19:26:31.314269] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.935 [2024-11-26 19:26:31.314493] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.935 [2024-11-26 19:26:31.314501] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.935 [2024-11-26 19:26:31.314509] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.935 [2024-11-26 19:26:31.314517] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.327151] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.327830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.327874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.935 [2024-11-26 19:26:31.327886] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.935 [2024-11-26 19:26:31.328126] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.935 [2024-11-26 19:26:31.328350] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.935 [2024-11-26 19:26:31.328359] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.935 [2024-11-26 19:26:31.328367] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.935 [2024-11-26 19:26:31.328374] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.341012] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.341601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.341620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.935 [2024-11-26 19:26:31.341628] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.935 [2024-11-26 19:26:31.341849] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.935 [2024-11-26 19:26:31.342081] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.935 [2024-11-26 19:26:31.342090] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.935 [2024-11-26 19:26:31.342097] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.935 [2024-11-26 19:26:31.342104] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.354938] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.355558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.355596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.935 [2024-11-26 19:26:31.355607] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.935 [2024-11-26 19:26:31.355846] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.935 [2024-11-26 19:26:31.356081] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.935 [2024-11-26 19:26:31.356091] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.935 [2024-11-26 19:26:31.356099] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.935 [2024-11-26 19:26:31.356106] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.368941] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.369617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.369654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.935 [2024-11-26 19:26:31.369665] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.935 [2024-11-26 19:26:31.369920] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.935 [2024-11-26 19:26:31.370146] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.935 [2024-11-26 19:26:31.370155] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.935 [2024-11-26 19:26:31.370163] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.935 [2024-11-26 19:26:31.370171] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.382806] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.383361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.383380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.935 [2024-11-26 19:26:31.383388] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.935 [2024-11-26 19:26:31.383608] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.935 [2024-11-26 19:26:31.383828] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.935 [2024-11-26 19:26:31.383837] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.935 [2024-11-26 19:26:31.383849] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.935 [2024-11-26 19:26:31.383856] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.935 [2024-11-26 19:26:31.396686] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.935 [2024-11-26 19:26:31.397143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.935 [2024-11-26 19:26:31.397161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.397168] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.397388] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.397608] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.397616] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.397623] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.397629] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.410665] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.411223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.936 [2024-11-26 19:26:31.411239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.411247] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.411466] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.411687] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.411695] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.411702] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.411709] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.424537] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.425165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.936 [2024-11-26 19:26:31.425203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.425214] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.425454] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.425678] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.425687] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.425695] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.425703] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.438550] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.439216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.936 [2024-11-26 19:26:31.439253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.439264] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.439504] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.439728] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.439738] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.439747] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.439755] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.452398] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.452995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.936 [2024-11-26 19:26:31.453033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.453046] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.453288] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.453512] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.453520] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.453528] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.453536] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.466394] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.467085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.936 [2024-11-26 19:26:31.467123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.467135] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.467376] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.467600] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.467609] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.467617] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.467625] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.480300] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.480961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.936 [2024-11-26 19:26:31.480998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.481014] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.481254] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.481478] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.481487] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.481494] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.481502] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.494145] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.494675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.936 [2024-11-26 19:26:31.494713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.936 [2024-11-26 19:26:31.494724] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.936 [2024-11-26 19:26:31.494974] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.936 [2024-11-26 19:26:31.495199] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.936 [2024-11-26 19:26:31.495208] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.936 [2024-11-26 19:26:31.495216] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.936 [2024-11-26 19:26:31.495223] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.936 [2024-11-26 19:26:31.508059] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.936 [2024-11-26 19:26:31.508730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.937 [2024-11-26 19:26:31.508768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.937 [2024-11-26 19:26:31.508779] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.937 [2024-11-26 19:26:31.509029] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.937 [2024-11-26 19:26:31.509254] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.937 [2024-11-26 19:26:31.509262] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.937 [2024-11-26 19:26:31.509271] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.937 [2024-11-26 19:26:31.509279] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.937 [2024-11-26 19:26:31.521921] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.937 [2024-11-26 19:26:31.522602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.937 [2024-11-26 19:26:31.522639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.937 [2024-11-26 19:26:31.522650] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.937 [2024-11-26 19:26:31.522900] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.937 [2024-11-26 19:26:31.523130] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.937 [2024-11-26 19:26:31.523139] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.937 [2024-11-26 19:26:31.523147] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.937 [2024-11-26 19:26:31.523155] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.937 [2024-11-26 19:26:31.535782] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.937 [2024-11-26 19:26:31.536353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.937 [2024-11-26 19:26:31.536391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.937 [2024-11-26 19:26:31.536404] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.937 [2024-11-26 19:26:31.536645] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.937 [2024-11-26 19:26:31.536880] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.937 [2024-11-26 19:26:31.536889] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.937 [2024-11-26 19:26:31.536897] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.937 [2024-11-26 19:26:31.536905] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:18.937 [2024-11-26 19:26:31.549737] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:18.937 [2024-11-26 19:26:31.550431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:18.937 [2024-11-26 19:26:31.550468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:18.937 [2024-11-26 19:26:31.550479] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:18.937 [2024-11-26 19:26:31.550720] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:18.937 [2024-11-26 19:26:31.550956] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:18.937 [2024-11-26 19:26:31.550967] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:18.937 [2024-11-26 19:26:31.550975] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:18.937 [2024-11-26 19:26:31.550983] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.200 [2024-11-26 19:26:31.563620] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.200 [2024-11-26 19:26:31.564298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.200 [2024-11-26 19:26:31.564336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.200 [2024-11-26 19:26:31.564347] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.200 [2024-11-26 19:26:31.564587] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.200 [2024-11-26 19:26:31.564810] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.200 [2024-11-26 19:26:31.564819] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.200 [2024-11-26 19:26:31.564832] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.200 [2024-11-26 19:26:31.564840] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.200 [2024-11-26 19:26:31.577486] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.578170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.578207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.578218] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.578458] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.578682] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.578691] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.578699] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.201 [2024-11-26 19:26:31.578707] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.201 [2024-11-26 19:26:31.591354] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.592042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.592080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.592091] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.592331] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.592555] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.592564] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.592571] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.201 [2024-11-26 19:26:31.592579] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.201 [2024-11-26 19:26:31.605215] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.605842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.605886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.605898] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.606139] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.606363] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.606372] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.606379] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.201 [2024-11-26 19:26:31.606387] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.201 [2024-11-26 19:26:31.619234] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.619937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.619975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.619987] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.620228] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.620453] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.620461] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.620469] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.201 [2024-11-26 19:26:31.620477] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.201 [2024-11-26 19:26:31.633117] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.633777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.633814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.633825] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.634074] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.634300] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.634309] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.634317] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.201 [2024-11-26 19:26:31.634324] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.201 [2024-11-26 19:26:31.646950] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.647542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.647561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.647569] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.647790] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.648017] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.648026] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.648034] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.201 [2024-11-26 19:26:31.648040] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.201 [2024-11-26 19:26:31.660866] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.661536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.661573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.661594] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.661834] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.662069] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.662079] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.662087] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.201 [2024-11-26 19:26:31.662094] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.201 [2024-11-26 19:26:31.674728] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.201 [2024-11-26 19:26:31.675370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.201 [2024-11-26 19:26:31.675408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.201 [2024-11-26 19:26:31.675419] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.201 [2024-11-26 19:26:31.675659] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.201 [2024-11-26 19:26:31.675893] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.201 [2024-11-26 19:26:31.675903] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.201 [2024-11-26 19:26:31.675910] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.675918] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.688547] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.689092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.689112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.689120] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.689341] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.689561] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.689569] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.689576] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.689583] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.702435] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.703069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.703107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.703118] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.703359] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.703589] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.703597] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.703606] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.703613] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.716362] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.716910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.716948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.716959] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.717199] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.717423] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.717432] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.717439] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.717448] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.730294] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.730934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.730972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.730983] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.731223] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.731447] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.731456] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.731464] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.731471] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.744316] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.745000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.745037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.745048] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.745289] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.745513] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.745521] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.745534] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.745541] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.758180] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.758883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.758921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.758934] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.759175] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.759400] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.759408] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.759416] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.759424] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.772067] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.772749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.772786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.772797] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.773047] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.773272] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.773280] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.773289] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.202 [2024-11-26 19:26:31.773297] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.202 [2024-11-26 19:26:31.785929] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.202 [2024-11-26 19:26:31.786604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.202 [2024-11-26 19:26:31.786642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.202 [2024-11-26 19:26:31.786653] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.202 [2024-11-26 19:26:31.786903] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.202 [2024-11-26 19:26:31.787128] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.202 [2024-11-26 19:26:31.787137] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.202 [2024-11-26 19:26:31.787144] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.203 [2024-11-26 19:26:31.787152] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.203 [2024-11-26 19:26:31.799775] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.203 [2024-11-26 19:26:31.800309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.203 [2024-11-26 19:26:31.800346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.203 [2024-11-26 19:26:31.800357] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.203 [2024-11-26 19:26:31.800597] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.203 [2024-11-26 19:26:31.800821] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.203 [2024-11-26 19:26:31.800829] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.203 [2024-11-26 19:26:31.800837] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.203 [2024-11-26 19:26:31.800845] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.203 [2024-11-26 19:26:31.813684] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.203 [2024-11-26 19:26:31.814312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.203 [2024-11-26 19:26:31.814350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.203 [2024-11-26 19:26:31.814361] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.203 [2024-11-26 19:26:31.814601] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.203 [2024-11-26 19:26:31.814825] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.203 [2024-11-26 19:26:31.814834] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.203 [2024-11-26 19:26:31.814842] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.203 [2024-11-26 19:26:31.814849] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.466 [2024-11-26 19:26:31.827706] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.466 [2024-11-26 19:26:31.828398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.466 [2024-11-26 19:26:31.828435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.466 [2024-11-26 19:26:31.828445] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.466 [2024-11-26 19:26:31.828685] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.466 [2024-11-26 19:26:31.828919] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.466 [2024-11-26 19:26:31.828929] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.466 [2024-11-26 19:26:31.828937] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.466 [2024-11-26 19:26:31.828944] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.466 [2024-11-26 19:26:31.841573] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.466 [2024-11-26 19:26:31.842139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.466 [2024-11-26 19:26:31.842176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.466 [2024-11-26 19:26:31.842191] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.466 [2024-11-26 19:26:31.842432] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.466 [2024-11-26 19:26:31.842656] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.466 [2024-11-26 19:26:31.842664] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.466 [2024-11-26 19:26:31.842672] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.466 [2024-11-26 19:26:31.842680] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.466 [2024-11-26 19:26:31.855523] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.466 [2024-11-26 19:26:31.856155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.466 [2024-11-26 19:26:31.856193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.466 [2024-11-26 19:26:31.856204] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.466 [2024-11-26 19:26:31.856444] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.466 [2024-11-26 19:26:31.856668] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.466 [2024-11-26 19:26:31.856676] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.466 [2024-11-26 19:26:31.856684] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.466 [2024-11-26 19:26:31.856692] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.466 [2024-11-26 19:26:31.869537] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.466 [2024-11-26 19:26:31.870216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.466 [2024-11-26 19:26:31.870254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.466 [2024-11-26 19:26:31.870266] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.466 [2024-11-26 19:26:31.870506] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.466 [2024-11-26 19:26:31.870730] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.466 [2024-11-26 19:26:31.870738] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.466 [2024-11-26 19:26:31.870746] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.466 [2024-11-26 19:26:31.870754] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.466 [2024-11-26 19:26:31.883399] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.466 [2024-11-26 19:26:31.884068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.466 [2024-11-26 19:26:31.884106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.466 [2024-11-26 19:26:31.884117] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.466 [2024-11-26 19:26:31.884357] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.466 [2024-11-26 19:26:31.884597] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.466 [2024-11-26 19:26:31.884607] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.466 [2024-11-26 19:26:31.884615] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.466 [2024-11-26 19:26:31.884622] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.897261] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.897899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.897936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.897947] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.898187] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.898411] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.898420] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.898428] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.898436] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.911278] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.911869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.911910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.911921] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.912161] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.912385] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.912394] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.912402] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.912411] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.925253] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.925810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.925829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.925837] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.926065] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.926286] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.926294] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.926305] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.926312] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.939132] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.939657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.939674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.939681] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.939908] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.940129] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.940137] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.940144] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.940151] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.953002] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.953534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.953551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.953559] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.953780] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.954007] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.954017] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.954025] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.954033] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.966964] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.967503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.967520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.967528] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.967748] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.967975] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.967985] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.967992] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.967999] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.980869] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.981446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.981463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.981470] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.981690] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.981917] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.981926] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.981934] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.981940] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:31.994788] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:31.995421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:31.995459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.467 [2024-11-26 19:26:31.995470] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.467 [2024-11-26 19:26:31.995710] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.467 [2024-11-26 19:26:31.995944] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.467 [2024-11-26 19:26:31.995953] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.467 [2024-11-26 19:26:31.995961] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.467 [2024-11-26 19:26:31.995969] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.467 [2024-11-26 19:26:32.008616] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.467 [2024-11-26 19:26:32.009269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.467 [2024-11-26 19:26:32.009307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.468 [2024-11-26 19:26:32.009318] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.468 [2024-11-26 19:26:32.009558] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.468 [2024-11-26 19:26:32.009782] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.468 [2024-11-26 19:26:32.009791] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.468 [2024-11-26 19:26:32.009798] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.468 [2024-11-26 19:26:32.009806] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.468 [2024-11-26 19:26:32.022461] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.468 [2024-11-26 19:26:32.023160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.468 [2024-11-26 19:26:32.023198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.468 [2024-11-26 19:26:32.023213] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.468 [2024-11-26 19:26:32.023453] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.468 [2024-11-26 19:26:32.023677] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.468 [2024-11-26 19:26:32.023686] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.468 [2024-11-26 19:26:32.023694] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.468 [2024-11-26 19:26:32.023702] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.468 [2024-11-26 19:26:32.036359] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.468 [2024-11-26 19:26:32.036967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.468 [2024-11-26 19:26:32.037006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.468 [2024-11-26 19:26:32.037018] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.468 [2024-11-26 19:26:32.037262] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.468 [2024-11-26 19:26:32.037485] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.468 [2024-11-26 19:26:32.037495] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.468 [2024-11-26 19:26:32.037503] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.468 [2024-11-26 19:26:32.037511] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.468 [2024-11-26 19:26:32.050360] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.468 [2024-11-26 19:26:32.050954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.468 [2024-11-26 19:26:32.050991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.468 [2024-11-26 19:26:32.051004] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.468 [2024-11-26 19:26:32.051247] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.468 [2024-11-26 19:26:32.051471] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.468 [2024-11-26 19:26:32.051480] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.468 [2024-11-26 19:26:32.051488] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.468 [2024-11-26 19:26:32.051496] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.468 [2024-11-26 19:26:32.064345] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.468 [2024-11-26 19:26:32.064921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.468 [2024-11-26 19:26:32.064941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.468 [2024-11-26 19:26:32.064949] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.468 [2024-11-26 19:26:32.065170] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.468 [2024-11-26 19:26:32.065395] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.468 [2024-11-26 19:26:32.065404] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.468 [2024-11-26 19:26:32.065411] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.468 [2024-11-26 19:26:32.065417] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.468 [2024-11-26 19:26:32.078274] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.468 [2024-11-26 19:26:32.078947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.468 [2024-11-26 19:26:32.078985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.468 [2024-11-26 19:26:32.078998] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.468 [2024-11-26 19:26:32.079239] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.468 [2024-11-26 19:26:32.079463] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.468 [2024-11-26 19:26:32.079472] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.468 [2024-11-26 19:26:32.079480] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.468 [2024-11-26 19:26:32.079488] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.731 [2024-11-26 19:26:32.092161] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.731 [2024-11-26 19:26:32.092790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.731 [2024-11-26 19:26:32.092828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.731 [2024-11-26 19:26:32.092839] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.731 [2024-11-26 19:26:32.093090] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.731 [2024-11-26 19:26:32.093315] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.731 [2024-11-26 19:26:32.093324] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.731 [2024-11-26 19:26:32.093332] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.731 [2024-11-26 19:26:32.093340] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.731 [2024-11-26 19:26:32.105990] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.731 [2024-11-26 19:26:32.106671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.731 [2024-11-26 19:26:32.106708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.731 [2024-11-26 19:26:32.106719] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.731 [2024-11-26 19:26:32.106969] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.731 [2024-11-26 19:26:32.107194] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.731 [2024-11-26 19:26:32.107202] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.731 [2024-11-26 19:26:32.107215] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.731 [2024-11-26 19:26:32.107223] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.731 [2024-11-26 19:26:32.119872] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.731 [2024-11-26 19:26:32.120421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.731 [2024-11-26 19:26:32.120440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.731 [2024-11-26 19:26:32.120448] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.731 [2024-11-26 19:26:32.120669] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.731 [2024-11-26 19:26:32.120896] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.731 [2024-11-26 19:26:32.120906] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.731 [2024-11-26 19:26:32.120913] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.731 [2024-11-26 19:26:32.120920] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.731 [2024-11-26 19:26:32.133768] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.731 [2024-11-26 19:26:32.134315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.731 [2024-11-26 19:26:32.134332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.731 [2024-11-26 19:26:32.134340] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.731 [2024-11-26 19:26:32.134560] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.731 [2024-11-26 19:26:32.134780] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.731 [2024-11-26 19:26:32.134788] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.731 [2024-11-26 19:26:32.134795] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.731 [2024-11-26 19:26:32.134801] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.731 [2024-11-26 19:26:32.147654] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.731 [2024-11-26 19:26:32.148193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.731 [2024-11-26 19:26:32.148210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.731 [2024-11-26 19:26:32.148217] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.731 [2024-11-26 19:26:32.148438] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.731 [2024-11-26 19:26:32.148657] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.148666] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.148673] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.148679] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.161536] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.162198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.162235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.162246] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.732 [2024-11-26 19:26:32.162486] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.732 [2024-11-26 19:26:32.162711] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.162719] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.162727] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.162735] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.175387] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.175997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.176035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.176046] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.732 [2024-11-26 19:26:32.176287] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.732 [2024-11-26 19:26:32.176511] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.176520] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.176527] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.176536] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.189407] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.189979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.190017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.190029] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.732 [2024-11-26 19:26:32.190273] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.732 [2024-11-26 19:26:32.190497] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.190505] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.190514] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.190522] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.203385] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.203908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.203929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.203942] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.732 [2024-11-26 19:26:32.204163] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.732 [2024-11-26 19:26:32.204384] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.204394] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.204404] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.204413] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.217266] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.217806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.217823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.217830] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.732 [2024-11-26 19:26:32.218058] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.732 [2024-11-26 19:26:32.218278] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.218286] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.218293] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.218300] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.231148] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.231687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.231703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.231710] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.732 [2024-11-26 19:26:32.231939] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.732 [2024-11-26 19:26:32.232159] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.232168] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.232175] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.232182] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.245039] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.245683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.245720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.245732] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.732 [2024-11-26 19:26:32.245982] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.732 [2024-11-26 19:26:32.246212] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.732 [2024-11-26 19:26:32.246221] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.732 [2024-11-26 19:26:32.246229] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.732 [2024-11-26 19:26:32.246237] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.732 [2024-11-26 19:26:32.258887] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.732 [2024-11-26 19:26:32.259446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.732 [2024-11-26 19:26:32.259466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.732 [2024-11-26 19:26:32.259474] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.733 [2024-11-26 19:26:32.259695] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.733 [2024-11-26 19:26:32.259923] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.733 [2024-11-26 19:26:32.259931] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.733 [2024-11-26 19:26:32.259939] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.733 [2024-11-26 19:26:32.259946] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.733 [2024-11-26 19:26:32.272809] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.733 [2024-11-26 19:26:32.273423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.733 [2024-11-26 19:26:32.273461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.733 [2024-11-26 19:26:32.273472] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.733 [2024-11-26 19:26:32.273712] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.733 [2024-11-26 19:26:32.273945] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.733 [2024-11-26 19:26:32.273955] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.733 [2024-11-26 19:26:32.273963] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.733 [2024-11-26 19:26:32.273971] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.733 7519.25 IOPS, 29.37 MiB/s [2024-11-26T18:26:32.358Z] [2024-11-26 19:26:32.286824] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.733 [2024-11-26 19:26:32.287367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.733 [2024-11-26 19:26:32.287405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.733 [2024-11-26 19:26:32.287418] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.733 [2024-11-26 19:26:32.287661] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.733 [2024-11-26 19:26:32.287895] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.733 [2024-11-26 19:26:32.287905] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.733 [2024-11-26 19:26:32.287917] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.733 [2024-11-26 19:26:32.287925] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.733 [2024-11-26 19:26:32.300787] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.733 [2024-11-26 19:26:32.301385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.733 [2024-11-26 19:26:32.301405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.733 [2024-11-26 19:26:32.301413] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.733 [2024-11-26 19:26:32.301634] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.733 [2024-11-26 19:26:32.301855] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.733 [2024-11-26 19:26:32.301871] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.733 [2024-11-26 19:26:32.301879] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.733 [2024-11-26 19:26:32.301886] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.733 [2024-11-26 19:26:32.314734] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.733 [2024-11-26 19:26:32.315272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.733 [2024-11-26 19:26:32.315289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.733 [2024-11-26 19:26:32.315296] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.733 [2024-11-26 19:26:32.315517] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.733 [2024-11-26 19:26:32.315736] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.733 [2024-11-26 19:26:32.315745] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.733 [2024-11-26 19:26:32.315752] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.733 [2024-11-26 19:26:32.315759] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.733 [2024-11-26 19:26:32.328611] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.733 [2024-11-26 19:26:32.329123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.733 [2024-11-26 19:26:32.329141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.733 [2024-11-26 19:26:32.329148] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.733 [2024-11-26 19:26:32.329369] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.733 [2024-11-26 19:26:32.329589] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.733 [2024-11-26 19:26:32.329597] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.733 [2024-11-26 19:26:32.329604] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.733 [2024-11-26 19:26:32.329611] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.733 [2024-11-26 19:26:32.342478] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.733 [2024-11-26 19:26:32.343012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.733 [2024-11-26 19:26:32.343030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.733 [2024-11-26 19:26:32.343037] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.733 [2024-11-26 19:26:32.343257] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.733 [2024-11-26 19:26:32.343477] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.733 [2024-11-26 19:26:32.343485] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.733 [2024-11-26 19:26:32.343493] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.733 [2024-11-26 19:26:32.343499] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.996 [2024-11-26 19:26:32.356348] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.996 [2024-11-26 19:26:32.357065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.996 [2024-11-26 19:26:32.357102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.996 [2024-11-26 19:26:32.357113] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.996 [2024-11-26 19:26:32.357353] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.996 [2024-11-26 19:26:32.357577] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.996 [2024-11-26 19:26:32.357586] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.996 [2024-11-26 19:26:32.357594] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.996 [2024-11-26 19:26:32.357602] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.996 [2024-11-26 19:26:32.370266] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.996 [2024-11-26 19:26:32.370820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.996 [2024-11-26 19:26:32.370840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.996 [2024-11-26 19:26:32.370848] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.996 [2024-11-26 19:26:32.371083] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.996 [2024-11-26 19:26:32.371305] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.996 [2024-11-26 19:26:32.371313] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.996 [2024-11-26 19:26:32.371321] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.996 [2024-11-26 19:26:32.371328] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.996 [2024-11-26 19:26:32.384172] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.996 [2024-11-26 19:26:32.384707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.996 [2024-11-26 19:26:32.384724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.996 [2024-11-26 19:26:32.384736] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.996 [2024-11-26 19:26:32.384963] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.996 [2024-11-26 19:26:32.385184] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.996 [2024-11-26 19:26:32.385192] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.996 [2024-11-26 19:26:32.385199] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.996 [2024-11-26 19:26:32.385207] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.996 [2024-11-26 19:26:32.398054] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.996 [2024-11-26 19:26:32.398631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.996 [2024-11-26 19:26:32.398648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.996 [2024-11-26 19:26:32.398656] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.996 [2024-11-26 19:26:32.398881] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.996 [2024-11-26 19:26:32.399101] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.996 [2024-11-26 19:26:32.399110] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.996 [2024-11-26 19:26:32.399117] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.399124] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.411968] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.412535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.412551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.412559] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.412778] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.997 [2024-11-26 19:26:32.413006] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.997 [2024-11-26 19:26:32.413015] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.997 [2024-11-26 19:26:32.413022] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.413029] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.425879] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.426410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.426427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.426434] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.426654] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.997 [2024-11-26 19:26:32.426883] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.997 [2024-11-26 19:26:32.426892] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.997 [2024-11-26 19:26:32.426899] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.426906] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.439746] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.440278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.440295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.440303] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.440522] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.997 [2024-11-26 19:26:32.440742] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.997 [2024-11-26 19:26:32.440750] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.997 [2024-11-26 19:26:32.440757] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.440764] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.453624] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.454073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.454090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.454098] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.454319] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.997 [2024-11-26 19:26:32.454539] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.997 [2024-11-26 19:26:32.454547] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.997 [2024-11-26 19:26:32.454554] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.454561] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.467625] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.468048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.468067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.468075] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.468296] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.997 [2024-11-26 19:26:32.468515] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.997 [2024-11-26 19:26:32.468524] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.997 [2024-11-26 19:26:32.468535] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.468542] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.481624] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.482164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.482181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.482189] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.482409] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.997 [2024-11-26 19:26:32.482629] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.997 [2024-11-26 19:26:32.482637] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.997 [2024-11-26 19:26:32.482644] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.482651] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.495510] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.496056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.496073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.496080] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.496300] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.997 [2024-11-26 19:26:32.496520] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.997 [2024-11-26 19:26:32.496529] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.997 [2024-11-26 19:26:32.496536] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.997 [2024-11-26 19:26:32.496542] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.997 [2024-11-26 19:26:32.509392] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.997 [2024-11-26 19:26:32.509932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.997 [2024-11-26 19:26:32.509949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.997 [2024-11-26 19:26:32.509957] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.997 [2024-11-26 19:26:32.510178] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.510397] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.510405] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.510413] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.510419] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.998 [2024-11-26 19:26:32.523278] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.998 [2024-11-26 19:26:32.523815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.998 [2024-11-26 19:26:32.523831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.998 [2024-11-26 19:26:32.523839] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.998 [2024-11-26 19:26:32.524064] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.524284] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.524299] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.524306] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.524313] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.998 [2024-11-26 19:26:32.537151] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.998 [2024-11-26 19:26:32.537682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.998 [2024-11-26 19:26:32.537698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.998 [2024-11-26 19:26:32.537706] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.998 [2024-11-26 19:26:32.537932] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.538153] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.538161] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.538168] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.538174] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.998 [2024-11-26 19:26:32.551019] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.998 [2024-11-26 19:26:32.551428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.998 [2024-11-26 19:26:32.551446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.998 [2024-11-26 19:26:32.551454] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.998 [2024-11-26 19:26:32.551674] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.551900] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.551908] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.551915] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.551922] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.998 [2024-11-26 19:26:32.564967] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.998 [2024-11-26 19:26:32.565503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.998 [2024-11-26 19:26:32.565519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.998 [2024-11-26 19:26:32.565530] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.998 [2024-11-26 19:26:32.565750] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.565976] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.565984] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.565991] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.565998] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.998 [2024-11-26 19:26:32.578845] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.998 [2024-11-26 19:26:32.579413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.998 [2024-11-26 19:26:32.579450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.998 [2024-11-26 19:26:32.579461] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.998 [2024-11-26 19:26:32.579702] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.579936] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.579945] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.579953] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.579961] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.998 [2024-11-26 19:26:32.592818] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.998 [2024-11-26 19:26:32.593373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.998 [2024-11-26 19:26:32.593393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.998 [2024-11-26 19:26:32.593401] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.998 [2024-11-26 19:26:32.593622] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.593842] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.593850] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.593857] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.593869] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:19.998 [2024-11-26 19:26:32.606701] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:19.998 [2024-11-26 19:26:32.607220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:19.998 [2024-11-26 19:26:32.607237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:19.998 [2024-11-26 19:26:32.607245] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:19.998 [2024-11-26 19:26:32.607465] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:19.998 [2024-11-26 19:26:32.607690] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:19.998 [2024-11-26 19:26:32.607698] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:19.998 [2024-11-26 19:26:32.607705] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:19.998 [2024-11-26 19:26:32.607712] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.261 [2024-11-26 19:26:32.620549] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.261 [2024-11-26 19:26:32.621164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.261 [2024-11-26 19:26:32.621202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.261 [2024-11-26 19:26:32.621215] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.261 [2024-11-26 19:26:32.621458] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.261 [2024-11-26 19:26:32.621682] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.261 [2024-11-26 19:26:32.621691] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.261 [2024-11-26 19:26:32.621699] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.261 [2024-11-26 19:26:32.621707] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.261 [2024-11-26 19:26:32.634556] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.261 [2024-11-26 19:26:32.635123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.261 [2024-11-26 19:26:32.635143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.261 [2024-11-26 19:26:32.635151] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.261 [2024-11-26 19:26:32.635372] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.261 [2024-11-26 19:26:32.635592] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.261 [2024-11-26 19:26:32.635601] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.261 [2024-11-26 19:26:32.635608] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.261 [2024-11-26 19:26:32.635615] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.261 [2024-11-26 19:26:32.648444] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.261 [2024-11-26 19:26:32.649159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.261 [2024-11-26 19:26:32.649197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.261 [2024-11-26 19:26:32.649208] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.261 [2024-11-26 19:26:32.649448] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.261 [2024-11-26 19:26:32.649672] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.261 [2024-11-26 19:26:32.649681] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.261 [2024-11-26 19:26:32.649693] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.261 [2024-11-26 19:26:32.649701] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.261 [2024-11-26 19:26:32.662342] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.261 [2024-11-26 19:26:32.662883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.261 [2024-11-26 19:26:32.662921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.261 [2024-11-26 19:26:32.662932] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.261 [2024-11-26 19:26:32.663172] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.261 [2024-11-26 19:26:32.663396] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.261 [2024-11-26 19:26:32.663405] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.261 [2024-11-26 19:26:32.663412] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.261 [2024-11-26 19:26:32.663420] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.261 [2024-11-26 19:26:32.676332] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.261 [2024-11-26 19:26:32.677094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.261 [2024-11-26 19:26:32.677133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.261 [2024-11-26 19:26:32.677144] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.261 [2024-11-26 19:26:32.677384] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.261 [2024-11-26 19:26:32.677608] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.261 [2024-11-26 19:26:32.677617] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.261 [2024-11-26 19:26:32.677625] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.261 [2024-11-26 19:26:32.677633] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.261 [2024-11-26 19:26:32.690282] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.261 [2024-11-26 19:26:32.690833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.261 [2024-11-26 19:26:32.690852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.261 [2024-11-26 19:26:32.690861] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.261 [2024-11-26 19:26:32.691088] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.261 [2024-11-26 19:26:32.691308] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.261 [2024-11-26 19:26:32.691317] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.261 [2024-11-26 19:26:32.691324] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.261 [2024-11-26 19:26:32.691331] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.261 [2024-11-26 19:26:32.704175] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.261 [2024-11-26 19:26:32.704844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.704889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.704901] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.705142] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.705366] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.705384] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.705393] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.705402] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.718042] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.262 [2024-11-26 19:26:32.718634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.718653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.718661] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.718887] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.719109] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.719118] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.719125] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.719133] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.731972] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.262 [2024-11-26 19:26:32.732507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.732524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.732532] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.732752] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.732978] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.732986] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.732994] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.733001] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.745827] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.262 [2024-11-26 19:26:32.746476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.746514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.746529] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.746770] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.747005] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.747015] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.747022] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.747030] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.759658] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.262 [2024-11-26 19:26:32.760299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.760336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.760347] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.760587] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.760811] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.760819] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.760827] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.760835] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.773689] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.262 [2024-11-26 19:26:32.774369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.774407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.774419] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.774662] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.774895] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.774904] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.774912] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.774920] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.787550] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.262 [2024-11-26 19:26:32.788221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.788259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.788270] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.788510] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.788739] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.788747] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.788755] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.788763] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.801402] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.262 [2024-11-26 19:26:32.801989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.262 [2024-11-26 19:26:32.802027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.262 [2024-11-26 19:26:32.802039] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.262 [2024-11-26 19:26:32.802282] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.262 [2024-11-26 19:26:32.802506] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.262 [2024-11-26 19:26:32.802515] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.262 [2024-11-26 19:26:32.802523] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.262 [2024-11-26 19:26:32.802531] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.262 [2024-11-26 19:26:32.815377] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.263 [2024-11-26 19:26:32.816045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.263 [2024-11-26 19:26:32.816083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.263 [2024-11-26 19:26:32.816094] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.263 [2024-11-26 19:26:32.816334] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.263 [2024-11-26 19:26:32.816558] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.263 [2024-11-26 19:26:32.816566] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.263 [2024-11-26 19:26:32.816574] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.263 [2024-11-26 19:26:32.816582] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.263 [2024-11-26 19:26:32.829222] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.263 [2024-11-26 19:26:32.829890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.263 [2024-11-26 19:26:32.829928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.263 [2024-11-26 19:26:32.829938] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.263 [2024-11-26 19:26:32.830178] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.263 [2024-11-26 19:26:32.830402] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.263 [2024-11-26 19:26:32.830410] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.263 [2024-11-26 19:26:32.830423] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.263 [2024-11-26 19:26:32.830431] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.263 [2024-11-26 19:26:32.843067] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.263 [2024-11-26 19:26:32.843616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.263 [2024-11-26 19:26:32.843636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.263 [2024-11-26 19:26:32.843644] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.263 [2024-11-26 19:26:32.843872] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.263 [2024-11-26 19:26:32.844093] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.263 [2024-11-26 19:26:32.844102] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.263 [2024-11-26 19:26:32.844109] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.263 [2024-11-26 19:26:32.844116] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.263 [2024-11-26 19:26:32.856935] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.263 [2024-11-26 19:26:32.857563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.263 [2024-11-26 19:26:32.857601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.263 [2024-11-26 19:26:32.857613] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.263 [2024-11-26 19:26:32.857856] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.263 [2024-11-26 19:26:32.858093] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.263 [2024-11-26 19:26:32.858101] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.263 [2024-11-26 19:26:32.858109] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.263 [2024-11-26 19:26:32.858117] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.263 [2024-11-26 19:26:32.870954] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.263 [2024-11-26 19:26:32.871584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.263 [2024-11-26 19:26:32.871622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.263 [2024-11-26 19:26:32.871633] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.263 [2024-11-26 19:26:32.871890] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.263 [2024-11-26 19:26:32.872117] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.263 [2024-11-26 19:26:32.872125] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.263 [2024-11-26 19:26:32.872133] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.263 [2024-11-26 19:26:32.872140] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.525 [2024-11-26 19:26:32.884976] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.525 [2024-11-26 19:26:32.885522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.525 [2024-11-26 19:26:32.885541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.525 [2024-11-26 19:26:32.885549] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.525 [2024-11-26 19:26:32.885770] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.525 [2024-11-26 19:26:32.885997] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.525 [2024-11-26 19:26:32.886006] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.525 [2024-11-26 19:26:32.886013] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.525 [2024-11-26 19:26:32.886020] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.525 [2024-11-26 19:26:32.898853] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.525 [2024-11-26 19:26:32.899435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.525 [2024-11-26 19:26:32.899451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.525 [2024-11-26 19:26:32.899459] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.525 [2024-11-26 19:26:32.899679] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.525 [2024-11-26 19:26:32.899904] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.525 [2024-11-26 19:26:32.899912] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.525 [2024-11-26 19:26:32.899919] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.525 [2024-11-26 19:26:32.899926] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:32.912757] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:32.913294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:32.913311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:32.913319] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:32.913539] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:32.913759] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:32.913767] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.526 [2024-11-26 19:26:32.913774] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.526 [2024-11-26 19:26:32.913780] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:32.926610] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:32.927148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:32.927165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:32.927177] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:32.927397] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:32.927617] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:32.927625] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.526 [2024-11-26 19:26:32.927632] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.526 [2024-11-26 19:26:32.927639] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:32.940469] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:32.941105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:32.941143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:32.941155] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:32.941396] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:32.941621] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:32.941630] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.526 [2024-11-26 19:26:32.941638] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.526 [2024-11-26 19:26:32.941646] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:32.954284] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:32.954991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:32.955029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:32.955040] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:32.955282] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:32.955506] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:32.955515] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.526 [2024-11-26 19:26:32.955523] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.526 [2024-11-26 19:26:32.955531] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:32.968169] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:32.968886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:32.968924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:32.968935] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:32.969175] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:32.969404] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:32.969412] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.526 [2024-11-26 19:26:32.969420] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.526 [2024-11-26 19:26:32.969428] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:32.982076] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:32.982731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:32.982768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:32.982779] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:32.983029] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:32.983255] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:32.983263] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.526 [2024-11-26 19:26:32.983271] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.526 [2024-11-26 19:26:32.983279] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:32.995999] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:32.996634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:32.996672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:32.996682] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:32.996931] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:32.997158] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:32.997166] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.526 [2024-11-26 19:26:32.997174] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.526 [2024-11-26 19:26:32.997182] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.526 [2024-11-26 19:26:33.010025] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.526 [2024-11-26 19:26:33.010700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.526 [2024-11-26 19:26:33.010739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.526 [2024-11-26 19:26:33.010750] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.526 [2024-11-26 19:26:33.011000] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.526 [2024-11-26 19:26:33.011225] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.526 [2024-11-26 19:26:33.011234] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.011250] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.011258] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.023889] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.024564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.024601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.024612] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.024852] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.527 [2024-11-26 19:26:33.025086] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.527 [2024-11-26 19:26:33.025095] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.025104] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.025112] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.037747] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.038430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.038468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.038479] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.038719] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.527 [2024-11-26 19:26:33.038952] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.527 [2024-11-26 19:26:33.038962] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.038970] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.038978] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.051596] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.052249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.052287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.052298] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.052539] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.527 [2024-11-26 19:26:33.052762] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.527 [2024-11-26 19:26:33.052771] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.052779] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.052787] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.065430] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.066026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.066046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.066054] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.066275] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.527 [2024-11-26 19:26:33.066495] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.527 [2024-11-26 19:26:33.066503] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.066511] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.066518] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.079363] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.079898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.079915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.079923] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.080143] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.527 [2024-11-26 19:26:33.080363] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.527 [2024-11-26 19:26:33.080371] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.080378] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.080385] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.093221] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.093889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.093927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.093939] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.094181] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.527 [2024-11-26 19:26:33.094405] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.527 [2024-11-26 19:26:33.094413] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.094421] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.094429] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.107074] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.107640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.107677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.107694] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.107944] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.527 [2024-11-26 19:26:33.108169] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.527 [2024-11-26 19:26:33.108178] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.527 [2024-11-26 19:26:33.108186] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.527 [2024-11-26 19:26:33.108194] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.527 [2024-11-26 19:26:33.121033] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.527 [2024-11-26 19:26:33.121710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.527 [2024-11-26 19:26:33.121748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.527 [2024-11-26 19:26:33.121759] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.527 [2024-11-26 19:26:33.122009] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.528 [2024-11-26 19:26:33.122235] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.528 [2024-11-26 19:26:33.122243] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.528 [2024-11-26 19:26:33.122250] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.528 [2024-11-26 19:26:33.122258] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.528 [2024-11-26 19:26:33.134892] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.528 [2024-11-26 19:26:33.135579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.528 [2024-11-26 19:26:33.135617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.528 [2024-11-26 19:26:33.135629] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.528 [2024-11-26 19:26:33.135879] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.528 [2024-11-26 19:26:33.136105] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.528 [2024-11-26 19:26:33.136114] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.528 [2024-11-26 19:26:33.136121] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.528 [2024-11-26 19:26:33.136129] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.791 [2024-11-26 19:26:33.148754] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.791 [2024-11-26 19:26:33.149419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.791 [2024-11-26 19:26:33.149456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.791 [2024-11-26 19:26:33.149467] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.791 [2024-11-26 19:26:33.149707] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.791 [2024-11-26 19:26:33.149945] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.791 [2024-11-26 19:26:33.149955] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.791 [2024-11-26 19:26:33.149963] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.791 [2024-11-26 19:26:33.149971] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.791 [2024-11-26 19:26:33.162598] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.791 [2024-11-26 19:26:33.163250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.791 [2024-11-26 19:26:33.163287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.791 [2024-11-26 19:26:33.163298] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.791 [2024-11-26 19:26:33.163537] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.791 [2024-11-26 19:26:33.163761] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.791 [2024-11-26 19:26:33.163770] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.791 [2024-11-26 19:26:33.163778] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.791 [2024-11-26 19:26:33.163786] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.791 [2024-11-26 19:26:33.176433] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.791 [2024-11-26 19:26:33.177170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.791 [2024-11-26 19:26:33.177208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.791 [2024-11-26 19:26:33.177220] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.791 [2024-11-26 19:26:33.177460] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.791 [2024-11-26 19:26:33.177684] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.791 [2024-11-26 19:26:33.177693] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.791 [2024-11-26 19:26:33.177700] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.791 [2024-11-26 19:26:33.177708] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.791 [2024-11-26 19:26:33.190356] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.791 [2024-11-26 19:26:33.190968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.791 [2024-11-26 19:26:33.191006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.791 [2024-11-26 19:26:33.191018] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.791 [2024-11-26 19:26:33.191262] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.791 [2024-11-26 19:26:33.191487] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.791 [2024-11-26 19:26:33.191496] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.791 [2024-11-26 19:26:33.191509] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.791 [2024-11-26 19:26:33.191517] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.791 [2024-11-26 19:26:33.204375] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.791 [2024-11-26 19:26:33.204971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.791 [2024-11-26 19:26:33.205010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.791 [2024-11-26 19:26:33.205022] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.791 [2024-11-26 19:26:33.205265] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.791 [2024-11-26 19:26:33.205489] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.791 [2024-11-26 19:26:33.205498] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.791 [2024-11-26 19:26:33.205505] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.791 [2024-11-26 19:26:33.205513] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.791 [2024-11-26 19:26:33.218360] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.791 [2024-11-26 19:26:33.219070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.791 [2024-11-26 19:26:33.219107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.791 [2024-11-26 19:26:33.219118] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.791 [2024-11-26 19:26:33.219358] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.791 [2024-11-26 19:26:33.219583] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.791 [2024-11-26 19:26:33.219591] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.791 [2024-11-26 19:26:33.219599] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.791 [2024-11-26 19:26:33.219607] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.791 [2024-11-26 19:26:33.232251] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.791 [2024-11-26 19:26:33.232945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.791 [2024-11-26 19:26:33.232983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.232994] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.233234] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.233458] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.233466] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.233474] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.233482] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 [2024-11-26 19:26:33.246170] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.792 [2024-11-26 19:26:33.246733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.792 [2024-11-26 19:26:33.246752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.246760] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.246986] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.247207] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.247214] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.247221] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.247229] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 [2024-11-26 19:26:33.260059] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.792 [2024-11-26 19:26:33.260680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.792 [2024-11-26 19:26:33.260718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.260729] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.260977] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.261202] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.261212] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.261220] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.261227] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 [2024-11-26 19:26:33.274086] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.792 [2024-11-26 19:26:33.274672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.792 [2024-11-26 19:26:33.274691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.274700] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.274930] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.275151] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.275159] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.275166] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.275173] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 6015.40 IOPS, 23.50 MiB/s [2024-11-26T18:26:33.417Z] [2024-11-26 19:26:33.287974] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.792 [2024-11-26 19:26:33.288574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.792 [2024-11-26 19:26:33.288612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.288627] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.288875] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.289100] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.289108] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.289117] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.289125] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 [2024-11-26 19:26:33.301972] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.792 [2024-11-26 19:26:33.302610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.792 [2024-11-26 19:26:33.302647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.302658] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.302906] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.303131] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.303140] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.303148] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.303156] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 [2024-11-26 19:26:33.315786] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.792 [2024-11-26 19:26:33.316452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.792 [2024-11-26 19:26:33.316490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.316503] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.316744] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.316978] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.316988] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.316996] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.317004] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 [2024-11-26 19:26:33.329632] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.792 [2024-11-26 19:26:33.330305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.792 [2024-11-26 19:26:33.330343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.792 [2024-11-26 19:26:33.330354] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.792 [2024-11-26 19:26:33.330593] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.792 [2024-11-26 19:26:33.330822] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.792 [2024-11-26 19:26:33.330831] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.792 [2024-11-26 19:26:33.330839] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.792 [2024-11-26 19:26:33.330846] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.792 [2024-11-26 19:26:33.343490] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.793 [2024-11-26 19:26:33.344141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.793 [2024-11-26 19:26:33.344179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.793 [2024-11-26 19:26:33.344190] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.793 [2024-11-26 19:26:33.344430] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.793 [2024-11-26 19:26:33.344654] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.793 [2024-11-26 19:26:33.344663] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.793 [2024-11-26 19:26:33.344670] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.793 [2024-11-26 19:26:33.344678] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.793 [2024-11-26 19:26:33.357315] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.793 [2024-11-26 19:26:33.357937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.793 [2024-11-26 19:26:33.357974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.793 [2024-11-26 19:26:33.357986] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.793 [2024-11-26 19:26:33.358230] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.793 [2024-11-26 19:26:33.358454] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.793 [2024-11-26 19:26:33.358462] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.793 [2024-11-26 19:26:33.358471] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.793 [2024-11-26 19:26:33.358478] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.793 [2024-11-26 19:26:33.371326] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.793 [2024-11-26 19:26:33.371963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.793 [2024-11-26 19:26:33.372001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.793 [2024-11-26 19:26:33.372014] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.793 [2024-11-26 19:26:33.372257] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.793 [2024-11-26 19:26:33.372481] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.793 [2024-11-26 19:26:33.372489] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.793 [2024-11-26 19:26:33.372502] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.793 [2024-11-26 19:26:33.372510] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.793 [2024-11-26 19:26:33.385159] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.793 [2024-11-26 19:26:33.385750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.793 [2024-11-26 19:26:33.385769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.793 [2024-11-26 19:26:33.385777] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.793 [2024-11-26 19:26:33.386004] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.793 [2024-11-26 19:26:33.386225] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.793 [2024-11-26 19:26:33.386233] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.793 [2024-11-26 19:26:33.386240] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.793 [2024-11-26 19:26:33.386247] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:20.793 [2024-11-26 19:26:33.399081] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:20.793 [2024-11-26 19:26:33.399744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:20.793 [2024-11-26 19:26:33.399782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:20.793 [2024-11-26 19:26:33.399793] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:20.793 [2024-11-26 19:26:33.400042] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:20.793 [2024-11-26 19:26:33.400267] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:20.793 [2024-11-26 19:26:33.400275] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:20.793 [2024-11-26 19:26:33.400283] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:20.793 [2024-11-26 19:26:33.400291] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.056 [2024-11-26 19:26:33.412917] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.056 [2024-11-26 19:26:33.413570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.056 [2024-11-26 19:26:33.413607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.056 [2024-11-26 19:26:33.413618] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.056 [2024-11-26 19:26:33.413858] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.056 [2024-11-26 19:26:33.414093] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.056 [2024-11-26 19:26:33.414102] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.056 [2024-11-26 19:26:33.414110] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.056 [2024-11-26 19:26:33.414118] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.056 [2024-11-26 19:26:33.426748] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.056 [2024-11-26 19:26:33.427414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.056 [2024-11-26 19:26:33.427452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.056 [2024-11-26 19:26:33.427464] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.056 [2024-11-26 19:26:33.427704] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.056 [2024-11-26 19:26:33.427937] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.056 [2024-11-26 19:26:33.427947] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.056 [2024-11-26 19:26:33.427954] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.056 [2024-11-26 19:26:33.427963] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.056 [2024-11-26 19:26:33.440582] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.056 [2024-11-26 19:26:33.441246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.056 [2024-11-26 19:26:33.441284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.056 [2024-11-26 19:26:33.441295] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.056 [2024-11-26 19:26:33.441535] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.056 [2024-11-26 19:26:33.441759] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.056 [2024-11-26 19:26:33.441769] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.056 [2024-11-26 19:26:33.441778] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.056 [2024-11-26 19:26:33.441787] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.056 [2024-11-26 19:26:33.454423] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.056 [2024-11-26 19:26:33.455179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.056 [2024-11-26 19:26:33.455217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.056 [2024-11-26 19:26:33.455229] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.056 [2024-11-26 19:26:33.455470] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.056 [2024-11-26 19:26:33.455695] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.056 [2024-11-26 19:26:33.455705] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.056 [2024-11-26 19:26:33.455713] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.056 [2024-11-26 19:26:33.455722] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.056 [2024-11-26 19:26:33.468357] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.056 [2024-11-26 19:26:33.468970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.056 [2024-11-26 19:26:33.469008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.056 [2024-11-26 19:26:33.469025] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.056 [2024-11-26 19:26:33.469268] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.056 [2024-11-26 19:26:33.469492] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.056 [2024-11-26 19:26:33.469500] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.056 [2024-11-26 19:26:33.469508] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.056 [2024-11-26 19:26:33.469516] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.056 [2024-11-26 19:26:33.482371] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.056 [2024-11-26 19:26:33.482807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.056 [2024-11-26 19:26:33.482827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.056 [2024-11-26 19:26:33.482835] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.056 [2024-11-26 19:26:33.483067] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.056 [2024-11-26 19:26:33.483288] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.056 [2024-11-26 19:26:33.483296] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.056 [2024-11-26 19:26:33.483303] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.056 [2024-11-26 19:26:33.483310] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.056 [2024-11-26 19:26:33.496354] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.497069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.497106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.497118] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.497358] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.497582] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.057 [2024-11-26 19:26:33.497590] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.057 [2024-11-26 19:26:33.497598] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.057 [2024-11-26 19:26:33.497606] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.057 [2024-11-26 19:26:33.510242] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.510920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.510958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.510971] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.511212] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.511441] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.057 [2024-11-26 19:26:33.511450] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.057 [2024-11-26 19:26:33.511458] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.057 [2024-11-26 19:26:33.511465] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.057 [2024-11-26 19:26:33.524101] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.524647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.524685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.524697] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.524949] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.525175] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.057 [2024-11-26 19:26:33.525183] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.057 [2024-11-26 19:26:33.525191] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.057 [2024-11-26 19:26:33.525199] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.057 [2024-11-26 19:26:33.538045] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.538723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.538762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.538773] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.539022] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.539247] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.057 [2024-11-26 19:26:33.539255] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.057 [2024-11-26 19:26:33.539263] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.057 [2024-11-26 19:26:33.539271] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.057 [2024-11-26 19:26:33.551899] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.552497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.552534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.552545] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.552785] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.553020] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.057 [2024-11-26 19:26:33.553030] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.057 [2024-11-26 19:26:33.553043] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.057 [2024-11-26 19:26:33.553051] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.057 [2024-11-26 19:26:33.565891] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.566556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.566594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.566605] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.566845] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.567079] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.057 [2024-11-26 19:26:33.567089] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.057 [2024-11-26 19:26:33.567097] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.057 [2024-11-26 19:26:33.567105] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.057 [2024-11-26 19:26:33.579747] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.580391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.580429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.580440] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.580680] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.580914] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.057 [2024-11-26 19:26:33.580923] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.057 [2024-11-26 19:26:33.580931] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.057 [2024-11-26 19:26:33.580939] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.057 [2024-11-26 19:26:33.593788] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.057 [2024-11-26 19:26:33.594430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.057 [2024-11-26 19:26:33.594468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.057 [2024-11-26 19:26:33.594479] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.057 [2024-11-26 19:26:33.594719] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.057 [2024-11-26 19:26:33.594952] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.058 [2024-11-26 19:26:33.594962] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.058 [2024-11-26 19:26:33.594970] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.058 [2024-11-26 19:26:33.594977] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.058 [2024-11-26 19:26:33.607810] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.058 [2024-11-26 19:26:33.608464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.058 [2024-11-26 19:26:33.608502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.058 [2024-11-26 19:26:33.608513] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.058 [2024-11-26 19:26:33.608753] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.058 [2024-11-26 19:26:33.608986] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.058 [2024-11-26 19:26:33.608996] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.058 [2024-11-26 19:26:33.609004] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.058 [2024-11-26 19:26:33.609012] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.058 [2024-11-26 19:26:33.621635] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.058 [2024-11-26 19:26:33.622205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.058 [2024-11-26 19:26:33.622224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.058 [2024-11-26 19:26:33.622232] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.058 [2024-11-26 19:26:33.622452] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.058 [2024-11-26 19:26:33.622673] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.058 [2024-11-26 19:26:33.622680] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.058 [2024-11-26 19:26:33.622688] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.058 [2024-11-26 19:26:33.622694] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.058 [2024-11-26 19:26:33.635523] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.058 [2024-11-26 19:26:33.636102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.058 [2024-11-26 19:26:33.636119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.058 [2024-11-26 19:26:33.636127] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.058 [2024-11-26 19:26:33.636346] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.058 [2024-11-26 19:26:33.636566] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.058 [2024-11-26 19:26:33.636573] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.058 [2024-11-26 19:26:33.636580] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.058 [2024-11-26 19:26:33.636587] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.058 [2024-11-26 19:26:33.649416] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.058 [2024-11-26 19:26:33.650063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.058 [2024-11-26 19:26:33.650101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.058 [2024-11-26 19:26:33.650116] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.058 [2024-11-26 19:26:33.650357] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.058 [2024-11-26 19:26:33.650581] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.058 [2024-11-26 19:26:33.650589] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.058 [2024-11-26 19:26:33.650597] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.058 [2024-11-26 19:26:33.650605] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.058 [2024-11-26 19:26:33.663243] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.058 [2024-11-26 19:26:33.663920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.058 [2024-11-26 19:26:33.663959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.058 [2024-11-26 19:26:33.663971] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.058 [2024-11-26 19:26:33.664212] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.058 [2024-11-26 19:26:33.664436] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.058 [2024-11-26 19:26:33.664444] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.058 [2024-11-26 19:26:33.664452] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.058 [2024-11-26 19:26:33.664460] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.058 [2024-11-26 19:26:33.677111] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.058 [2024-11-26 19:26:33.677693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.058 [2024-11-26 19:26:33.677730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.058 [2024-11-26 19:26:33.677741] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.677991] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.678218] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.678227] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.678235] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.678242] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.691188] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.691882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.691920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.691931] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.692171] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.692405] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.692415] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.692424] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.692433] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.705060] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.705742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.705779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.705791] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.706040] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.706266] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.706275] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.706282] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.706290] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.718923] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.719606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.719643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.719654] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.719904] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.720130] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.720138] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.720146] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.720154] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.732784] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.733452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.733490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.733501] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.733741] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.733974] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.733983] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.733996] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.734004] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.746635] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.747301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.747339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.747350] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.747590] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.747814] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.747822] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.747830] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.747838] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.760483] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.761138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.761175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.761186] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.761426] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.761650] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.761658] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.761667] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.761675] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.774325] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.774911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.774930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.774939] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.775160] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.775380] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.775388] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.775396] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.775403] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.788243] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.788880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.788918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.788931] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.789174] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.789398] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.789406] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.789414] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.789422] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.802073] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.322 [2024-11-26 19:26:33.802662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.322 [2024-11-26 19:26:33.802681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.322 [2024-11-26 19:26:33.802689] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.322 [2024-11-26 19:26:33.802914] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.322 [2024-11-26 19:26:33.803135] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.322 [2024-11-26 19:26:33.803144] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.322 [2024-11-26 19:26:33.803151] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.322 [2024-11-26 19:26:33.803157] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.322 [2024-11-26 19:26:33.815991] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.816608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.816645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.816656] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.816904] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.817129] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.817138] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.817146] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.817153] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 [2024-11-26 19:26:33.830000] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.830623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.830661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.830677] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.830925] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.831150] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.831159] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.831167] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.831175] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 [2024-11-26 19:26:33.844022] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.844713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.844751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.844762] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.845011] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.845236] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.845245] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.845252] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.845260] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 [2024-11-26 19:26:33.857922] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.858552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.858589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.858602] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.858843] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.859077] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.859087] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.859095] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.859102] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 [2024-11-26 19:26:33.871943] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.872624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.872662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.872673] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.872926] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.873156] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.873165] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.873173] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.873180] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 [2024-11-26 19:26:33.885817] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.886412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.886432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.886439] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.886660] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.886885] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.886894] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.886901] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.886908] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 [2024-11-26 19:26:33.899745] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.900201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.900219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.900227] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.900447] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.900666] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.900675] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.900682] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.900689] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 [2024-11-26 19:26:33.913729] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 [2024-11-26 19:26:33.914277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.914293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.914301] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 [2024-11-26 19:26:33.914521] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.914741] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.914748] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.914759] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.914766] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh: line 35: 4056635 Killed "${NVMF_APP[@]}" "$@" 00:30:21.323 [2024-11-26 19:26:33.927599] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@36 -- # tgt_init 00:30:21.323 [2024-11-26 19:26:33.928232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.323 [2024-11-26 19:26:33.928271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.323 [2024-11-26 19:26:33.928282] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:30:21.323 [2024-11-26 19:26:33.928522] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.323 [2024-11-26 19:26:33.928746] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.323 [2024-11-26 19:26:33.928755] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.323 [2024-11-26 19:26:33.928763] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.323 [2024-11-26 19:26:33.928771] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@726 -- # xtrace_disable 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # nvmfpid=4058088 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # waitforlisten 4058088 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # '[' -z 4058088 ']' 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:21.323 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:21.324 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:21.324 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:21.324 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:21.324 19:26:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:21.324 [2024-11-26 19:26:33.941411] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.324 [2024-11-26 19:26:33.941979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.324 [2024-11-26 19:26:33.941998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.324 [2024-11-26 19:26:33.942006] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.324 [2024-11-26 19:26:33.942227] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.324 [2024-11-26 19:26:33.942448] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.324 [2024-11-26 19:26:33.942461] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.324 [2024-11-26 19:26:33.942469] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.324 [2024-11-26 19:26:33.942476] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.586 [2024-11-26 19:26:33.955316] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.586 [2024-11-26 19:26:33.955879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.586 [2024-11-26 19:26:33.955918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.586 [2024-11-26 19:26:33.955930] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.586 [2024-11-26 19:26:33.956174] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.586 [2024-11-26 19:26:33.956399] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.586 [2024-11-26 19:26:33.956408] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.586 [2024-11-26 19:26:33.956417] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.586 [2024-11-26 19:26:33.956426] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.586 [2024-11-26 19:26:33.969277] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.586 [2024-11-26 19:26:33.969884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.586 [2024-11-26 19:26:33.969921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.586 [2024-11-26 19:26:33.969933] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.586 [2024-11-26 19:26:33.970173] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.586 [2024-11-26 19:26:33.970397] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.586 [2024-11-26 19:26:33.970406] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.586 [2024-11-26 19:26:33.970414] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.586 [2024-11-26 19:26:33.970422] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.586 [2024-11-26 19:26:33.983289] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.586 [2024-11-26 19:26:33.983880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.586 [2024-11-26 19:26:33.983900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.586 [2024-11-26 19:26:33.983908] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.586 [2024-11-26 19:26:33.984129] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.586 [2024-11-26 19:26:33.984351] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.586 [2024-11-26 19:26:33.984359] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.586 [2024-11-26 19:26:33.984366] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.586 [2024-11-26 19:26:33.984378] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.586 [2024-11-26 19:26:33.991004] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:30:21.586 [2024-11-26 19:26:33.991050] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:21.586 [2024-11-26 19:26:33.997227] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.586 [2024-11-26 19:26:33.997921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.586 [2024-11-26 19:26:33.997959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.586 [2024-11-26 19:26:33.997971] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.586 [2024-11-26 19:26:33.998211] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.586 [2024-11-26 19:26:33.998436] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.586 [2024-11-26 19:26:33.998445] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.586 [2024-11-26 19:26:33.998453] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.586 [2024-11-26 19:26:33.998461] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.586 [2024-11-26 19:26:34.011111] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.586 [2024-11-26 19:26:34.011806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.586 [2024-11-26 19:26:34.011844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.586 [2024-11-26 19:26:34.011857] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.586 [2024-11-26 19:26:34.012108] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.586 [2024-11-26 19:26:34.012332] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.586 [2024-11-26 19:26:34.012342] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.586 [2024-11-26 19:26:34.012350] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.586 [2024-11-26 19:26:34.012358] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.586 [2024-11-26 19:26:34.025073] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.586 [2024-11-26 19:26:34.025720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.586 [2024-11-26 19:26:34.025758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.586 [2024-11-26 19:26:34.025769] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.586 [2024-11-26 19:26:34.026017] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.586 [2024-11-26 19:26:34.026242] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.586 [2024-11-26 19:26:34.026251] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.586 [2024-11-26 19:26:34.026259] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.026272] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.038920] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.039501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.039539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.039552] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.039793] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.587 [2024-11-26 19:26:34.040026] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.587 [2024-11-26 19:26:34.040037] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.587 [2024-11-26 19:26:34.040045] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.040053] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.052905] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.053607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.053644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.053656] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.053905] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.587 [2024-11-26 19:26:34.054131] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.587 [2024-11-26 19:26:34.054140] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.587 [2024-11-26 19:26:34.054148] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.054156] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.066790] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.067314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.067333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.067342] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.067562] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.587 [2024-11-26 19:26:34.067782] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.587 [2024-11-26 19:26:34.067791] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.587 [2024-11-26 19:26:34.067798] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.067805] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.080652] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.081187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.081225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.081236] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.081476] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.587 [2024-11-26 19:26:34.081701] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.587 [2024-11-26 19:26:34.081710] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.587 [2024-11-26 19:26:34.081718] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.081726] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.085597] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:30:21.587 [2024-11-26 19:26:34.094603] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.095185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.095205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.095214] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.095435] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.587 [2024-11-26 19:26:34.095656] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.587 [2024-11-26 19:26:34.095664] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.587 [2024-11-26 19:26:34.095671] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.095679] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.108526] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.109100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.109117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.109125] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.109347] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.587 [2024-11-26 19:26:34.109567] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.587 [2024-11-26 19:26:34.109582] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.587 [2024-11-26 19:26:34.109590] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.109597] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.114928] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:21.587 [2024-11-26 19:26:34.114951] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:21.587 [2024-11-26 19:26:34.114958] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:21.587 [2024-11-26 19:26:34.114966] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:21.587 [2024-11-26 19:26:34.114971] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:21.587 [2024-11-26 19:26:34.116237] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:30:21.587 [2024-11-26 19:26:34.116391] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:30:21.587 [2024-11-26 19:26:34.116394] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:30:21.587 [2024-11-26 19:26:34.122453] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.123184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.123225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.123237] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.123480] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.587 [2024-11-26 19:26:34.123705] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.587 [2024-11-26 19:26:34.123714] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.587 [2024-11-26 19:26:34.123722] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.587 [2024-11-26 19:26:34.123731] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.587 [2024-11-26 19:26:34.136383] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.587 [2024-11-26 19:26:34.137088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.587 [2024-11-26 19:26:34.137128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.587 [2024-11-26 19:26:34.137139] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.587 [2024-11-26 19:26:34.137381] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.588 [2024-11-26 19:26:34.137605] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.588 [2024-11-26 19:26:34.137614] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.588 [2024-11-26 19:26:34.137622] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.588 [2024-11-26 19:26:34.137630] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.588 [2024-11-26 19:26:34.150277] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.588 [2024-11-26 19:26:34.150977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.588 [2024-11-26 19:26:34.151016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.588 [2024-11-26 19:26:34.151027] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.588 [2024-11-26 19:26:34.151268] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.588 [2024-11-26 19:26:34.151493] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.588 [2024-11-26 19:26:34.151501] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.588 [2024-11-26 19:26:34.151516] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.588 [2024-11-26 19:26:34.151524] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.588 [2024-11-26 19:26:34.164168] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.588 [2024-11-26 19:26:34.164880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.588 [2024-11-26 19:26:34.164918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.588 [2024-11-26 19:26:34.164930] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.588 [2024-11-26 19:26:34.165172] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.588 [2024-11-26 19:26:34.165396] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.588 [2024-11-26 19:26:34.165405] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.588 [2024-11-26 19:26:34.165413] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.588 [2024-11-26 19:26:34.165422] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.588 [2024-11-26 19:26:34.178072] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.588 [2024-11-26 19:26:34.178774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.588 [2024-11-26 19:26:34.178812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.588 [2024-11-26 19:26:34.178824] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.588 [2024-11-26 19:26:34.179073] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.588 [2024-11-26 19:26:34.179298] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.588 [2024-11-26 19:26:34.179307] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.588 [2024-11-26 19:26:34.179315] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.588 [2024-11-26 19:26:34.179323] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.588 [2024-11-26 19:26:34.191958] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.588 [2024-11-26 19:26:34.192630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.588 [2024-11-26 19:26:34.192667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.588 [2024-11-26 19:26:34.192679] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.588 [2024-11-26 19:26:34.192927] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.588 [2024-11-26 19:26:34.193166] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.588 [2024-11-26 19:26:34.193176] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.588 [2024-11-26 19:26:34.193184] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.588 [2024-11-26 19:26:34.193192] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.588 [2024-11-26 19:26:34.205830] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.588 [2024-11-26 19:26:34.206538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.588 [2024-11-26 19:26:34.206577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.588 [2024-11-26 19:26:34.206589] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.588 [2024-11-26 19:26:34.206829] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.588 [2024-11-26 19:26:34.207064] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.588 [2024-11-26 19:26:34.207073] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.588 [2024-11-26 19:26:34.207081] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.588 [2024-11-26 19:26:34.207089] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 [2024-11-26 19:26:34.219729] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.851 [2024-11-26 19:26:34.220296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.851 [2024-11-26 19:26:34.220315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.851 [2024-11-26 19:26:34.220324] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.851 [2024-11-26 19:26:34.220545] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.851 [2024-11-26 19:26:34.220764] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.851 [2024-11-26 19:26:34.220773] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.851 [2024-11-26 19:26:34.220780] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.851 [2024-11-26 19:26:34.220787] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 [2024-11-26 19:26:34.233635] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.851 [2024-11-26 19:26:34.234166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.851 [2024-11-26 19:26:34.234204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.851 [2024-11-26 19:26:34.234215] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.851 [2024-11-26 19:26:34.234456] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.851 [2024-11-26 19:26:34.234680] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.851 [2024-11-26 19:26:34.234689] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.851 [2024-11-26 19:26:34.234697] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.851 [2024-11-26 19:26:34.234705] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 [2024-11-26 19:26:34.247555] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.851 [2024-11-26 19:26:34.248219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.851 [2024-11-26 19:26:34.248257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.851 [2024-11-26 19:26:34.248273] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.851 [2024-11-26 19:26:34.248514] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.851 [2024-11-26 19:26:34.248738] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.851 [2024-11-26 19:26:34.248748] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.851 [2024-11-26 19:26:34.248756] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.851 [2024-11-26 19:26:34.248765] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 [2024-11-26 19:26:34.261411] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.851 [2024-11-26 19:26:34.261849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.851 [2024-11-26 19:26:34.261874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.851 [2024-11-26 19:26:34.261882] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.851 [2024-11-26 19:26:34.262103] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.851 [2024-11-26 19:26:34.262324] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.851 [2024-11-26 19:26:34.262333] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.851 [2024-11-26 19:26:34.262340] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.851 [2024-11-26 19:26:34.262347] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 [2024-11-26 19:26:34.275404] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.851 [2024-11-26 19:26:34.275960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.851 [2024-11-26 19:26:34.275978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.851 [2024-11-26 19:26:34.275985] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.851 [2024-11-26 19:26:34.276206] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.851 [2024-11-26 19:26:34.276426] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.851 [2024-11-26 19:26:34.276434] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.851 [2024-11-26 19:26:34.276441] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.851 [2024-11-26 19:26:34.276447] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 5012.83 IOPS, 19.58 MiB/s [2024-11-26T18:26:34.476Z] [2024-11-26 19:26:34.289257] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.851 [2024-11-26 19:26:34.289848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.851 [2024-11-26 19:26:34.289869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.851 [2024-11-26 19:26:34.289878] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.851 [2024-11-26 19:26:34.290097] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.851 [2024-11-26 19:26:34.290322] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.851 [2024-11-26 19:26:34.290330] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.851 [2024-11-26 19:26:34.290337] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.851 [2024-11-26 19:26:34.290344] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 [2024-11-26 19:26:34.303182] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.851 [2024-11-26 19:26:34.303832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.851 [2024-11-26 19:26:34.303877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.851 [2024-11-26 19:26:34.303890] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.851 [2024-11-26 19:26:34.304130] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.851 [2024-11-26 19:26:34.304354] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.851 [2024-11-26 19:26:34.304364] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.851 [2024-11-26 19:26:34.304371] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.851 [2024-11-26 19:26:34.304380] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.851 [2024-11-26 19:26:34.317019] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.317622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.317642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.317650] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.317876] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.318097] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.318105] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.852 [2024-11-26 19:26:34.318112] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.852 [2024-11-26 19:26:34.318120] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.852 [2024-11-26 19:26:34.330955] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.331473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.331511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.331523] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.331763] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.331996] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.332005] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.852 [2024-11-26 19:26:34.332018] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.852 [2024-11-26 19:26:34.332026] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.852 [2024-11-26 19:26:34.344873] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.345584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.345622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.345633] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.345881] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.346106] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.346115] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.852 [2024-11-26 19:26:34.346123] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.852 [2024-11-26 19:26:34.346131] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.852 [2024-11-26 19:26:34.358758] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.359362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.359381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.359390] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.359610] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.359831] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.359839] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.852 [2024-11-26 19:26:34.359846] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.852 [2024-11-26 19:26:34.359853] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.852 [2024-11-26 19:26:34.372686] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.373332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.373371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.373382] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.373622] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.373847] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.373855] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.852 [2024-11-26 19:26:34.373871] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.852 [2024-11-26 19:26:34.373879] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.852 [2024-11-26 19:26:34.386505] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.387199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.387237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.387248] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.387489] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.387712] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.387722] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.852 [2024-11-26 19:26:34.387730] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.852 [2024-11-26 19:26:34.387737] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.852 [2024-11-26 19:26:34.400445] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.400994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.401034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.401045] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.401285] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.401510] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.401519] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.852 [2024-11-26 19:26:34.401527] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.852 [2024-11-26 19:26:34.401535] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.852 [2024-11-26 19:26:34.414387] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.852 [2024-11-26 19:26:34.414968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.852 [2024-11-26 19:26:34.415006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.852 [2024-11-26 19:26:34.415019] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.852 [2024-11-26 19:26:34.415263] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.852 [2024-11-26 19:26:34.415488] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.852 [2024-11-26 19:26:34.415496] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.853 [2024-11-26 19:26:34.415505] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.853 [2024-11-26 19:26:34.415513] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.853 [2024-11-26 19:26:34.428370] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.853 [2024-11-26 19:26:34.428990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.853 [2024-11-26 19:26:34.429028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.853 [2024-11-26 19:26:34.429045] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.853 [2024-11-26 19:26:34.429285] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.853 [2024-11-26 19:26:34.429509] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.853 [2024-11-26 19:26:34.429518] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.853 [2024-11-26 19:26:34.429526] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.853 [2024-11-26 19:26:34.429534] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.853 [2024-11-26 19:26:34.442383] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.853 [2024-11-26 19:26:34.443005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.853 [2024-11-26 19:26:34.443043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.853 [2024-11-26 19:26:34.443054] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.853 [2024-11-26 19:26:34.443295] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.853 [2024-11-26 19:26:34.443519] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.853 [2024-11-26 19:26:34.443529] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.853 [2024-11-26 19:26:34.443536] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.853 [2024-11-26 19:26:34.443545] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.853 [2024-11-26 19:26:34.456398] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.853 [2024-11-26 19:26:34.457117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.853 [2024-11-26 19:26:34.457156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.853 [2024-11-26 19:26:34.457167] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.853 [2024-11-26 19:26:34.457407] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.853 [2024-11-26 19:26:34.457632] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.853 [2024-11-26 19:26:34.457641] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.853 [2024-11-26 19:26:34.457648] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.853 [2024-11-26 19:26:34.457656] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:21.853 [2024-11-26 19:26:34.470295] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:21.853 [2024-11-26 19:26:34.470915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:21.853 [2024-11-26 19:26:34.470940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:21.853 [2024-11-26 19:26:34.470948] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:21.853 [2024-11-26 19:26:34.471174] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:21.853 [2024-11-26 19:26:34.471400] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:21.853 [2024-11-26 19:26:34.471408] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:21.853 [2024-11-26 19:26:34.471415] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:21.853 [2024-11-26 19:26:34.471423] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.114 [2024-11-26 19:26:34.484278] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.114 [2024-11-26 19:26:34.484969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.114 [2024-11-26 19:26:34.485007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.114 [2024-11-26 19:26:34.485018] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.114 [2024-11-26 19:26:34.485258] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.114 [2024-11-26 19:26:34.485482] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.114 [2024-11-26 19:26:34.485491] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.114 [2024-11-26 19:26:34.485500] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.114 [2024-11-26 19:26:34.485508] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.114 [2024-11-26 19:26:34.498161] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.114 [2024-11-26 19:26:34.498838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.114 [2024-11-26 19:26:34.498883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.114 [2024-11-26 19:26:34.498895] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.114 [2024-11-26 19:26:34.499135] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.114 [2024-11-26 19:26:34.499359] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.114 [2024-11-26 19:26:34.499367] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.114 [2024-11-26 19:26:34.499375] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.114 [2024-11-26 19:26:34.499383] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.114 [2024-11-26 19:26:34.512015] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.114 [2024-11-26 19:26:34.512474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.114 [2024-11-26 19:26:34.512512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.114 [2024-11-26 19:26:34.512524] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.512767] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.512998] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.513009] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.513022] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.513029] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.525868] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.526465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.526484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.526492] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.526713] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.526939] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.526947] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.526954] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.526961] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.539791] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.540454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.540491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.540502] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.540743] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.540975] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.540984] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.540992] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.541000] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.553625] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.554267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.554305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.554316] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.554556] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.554780] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.554789] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.554797] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.554805] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.567647] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.568328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.568367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.568378] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.568618] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.568842] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.568850] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.568858] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.568875] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.581513] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.581987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.582025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.582037] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.582281] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.582504] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.582513] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.582521] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.582528] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.595379] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.596005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.596043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.596054] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.596294] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.596518] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.596526] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.596534] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.596542] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.609381] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.610111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.610148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.610164] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.610404] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.610628] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.610638] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.610646] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.610653] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.623286] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.623833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.623878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.623891] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.624134] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.115 [2024-11-26 19:26:34.624358] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.115 [2024-11-26 19:26:34.624367] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.115 [2024-11-26 19:26:34.624375] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.115 [2024-11-26 19:26:34.624383] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.115 [2024-11-26 19:26:34.637234] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.115 [2024-11-26 19:26:34.637920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.115 [2024-11-26 19:26:34.637958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.115 [2024-11-26 19:26:34.637970] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.115 [2024-11-26 19:26:34.638211] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.638435] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.638443] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.638451] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.638458] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.116 [2024-11-26 19:26:34.651097] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.116 [2024-11-26 19:26:34.651790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.116 [2024-11-26 19:26:34.651828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.116 [2024-11-26 19:26:34.651839] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.116 [2024-11-26 19:26:34.652088] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.652318] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.652326] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.652334] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.652342] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.116 [2024-11-26 19:26:34.664972] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.116 [2024-11-26 19:26:34.665638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.116 [2024-11-26 19:26:34.665677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.116 [2024-11-26 19:26:34.665688] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.116 [2024-11-26 19:26:34.665937] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.666163] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.666172] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.666180] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.666187] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.116 [2024-11-26 19:26:34.678827] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.116 [2024-11-26 19:26:34.679442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.116 [2024-11-26 19:26:34.679480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.116 [2024-11-26 19:26:34.679491] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.116 [2024-11-26 19:26:34.679731] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.679963] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.679972] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.679980] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.679988] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.116 [2024-11-26 19:26:34.692826] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.116 [2024-11-26 19:26:34.693527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.116 [2024-11-26 19:26:34.693564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.116 [2024-11-26 19:26:34.693576] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.116 [2024-11-26 19:26:34.693816] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.694048] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.694058] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.694071] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.694080] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.116 [2024-11-26 19:26:34.706720] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.116 [2024-11-26 19:26:34.707393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.116 [2024-11-26 19:26:34.707432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.116 [2024-11-26 19:26:34.707443] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.116 [2024-11-26 19:26:34.707683] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.707917] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.707927] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.707936] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.707943] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.116 [2024-11-26 19:26:34.720661] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.116 [2024-11-26 19:26:34.721377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.116 [2024-11-26 19:26:34.721415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.116 [2024-11-26 19:26:34.721426] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.116 [2024-11-26 19:26:34.721667] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.721900] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.721921] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.721929] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.721937] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.116 [2024-11-26 19:26:34.734573] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.116 [2024-11-26 19:26:34.735120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.116 [2024-11-26 19:26:34.735158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.116 [2024-11-26 19:26:34.735170] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.116 [2024-11-26 19:26:34.735409] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.116 [2024-11-26 19:26:34.735633] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.116 [2024-11-26 19:26:34.735642] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.116 [2024-11-26 19:26:34.735650] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.116 [2024-11-26 19:26:34.735658] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.377 [2024-11-26 19:26:34.748500] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.377 [2024-11-26 19:26:34.749171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.377 [2024-11-26 19:26:34.749209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.377 [2024-11-26 19:26:34.749220] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.377 [2024-11-26 19:26:34.749460] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.377 [2024-11-26 19:26:34.749685] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.377 [2024-11-26 19:26:34.749693] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.377 [2024-11-26 19:26:34.749701] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.377 [2024-11-26 19:26:34.749709] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.377 [2024-11-26 19:26:34.762343] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.377 [2024-11-26 19:26:34.762930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.377 [2024-11-26 19:26:34.762968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.377 [2024-11-26 19:26:34.762981] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.377 [2024-11-26 19:26:34.763223] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.377 [2024-11-26 19:26:34.763447] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.377 [2024-11-26 19:26:34.763455] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.377 [2024-11-26 19:26:34.763463] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.377 [2024-11-26 19:26:34.763471] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.377 [2024-11-26 19:26:34.776322] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.377 [2024-11-26 19:26:34.777103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.377 [2024-11-26 19:26:34.777141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.377 [2024-11-26 19:26:34.777152] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.377 [2024-11-26 19:26:34.777393] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.377 [2024-11-26 19:26:34.777616] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.377 [2024-11-26 19:26:34.777625] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.377 [2024-11-26 19:26:34.777633] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.377 [2024-11-26 19:26:34.777641] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.377 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:22.377 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@868 -- # return 0 00:30:22.377 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:30:22.377 [2024-11-26 19:26:34.790273] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.377 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@732 -- # xtrace_disable 00:30:22.377 [2024-11-26 19:26:34.790591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.377 [2024-11-26 19:26:34.790615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.377 [2024-11-26 19:26:34.790624] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.377 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:22.377 [2024-11-26 19:26:34.790850] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.377 [2024-11-26 19:26:34.791079] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.377 [2024-11-26 19:26:34.791088] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.377 [2024-11-26 19:26:34.791095] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.377 [2024-11-26 19:26:34.791102] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.377 [2024-11-26 19:26:34.804156] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.377 [2024-11-26 19:26:34.804791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.377 [2024-11-26 19:26:34.804828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.377 [2024-11-26 19:26:34.804841] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.377 [2024-11-26 19:26:34.805094] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.377 [2024-11-26 19:26:34.805319] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.377 [2024-11-26 19:26:34.805327] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.377 [2024-11-26 19:26:34.805335] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.377 [2024-11-26 19:26:34.805343] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.377 [2024-11-26 19:26:34.817976] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.377 [2024-11-26 19:26:34.818676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.377 [2024-11-26 19:26:34.818715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.377 [2024-11-26 19:26:34.818726] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.377 [2024-11-26 19:26:34.818976] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.378 [2024-11-26 19:26:34.819201] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.378 [2024-11-26 19:26:34.819210] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.378 [2024-11-26 19:26:34.819218] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.378 [2024-11-26 19:26:34.819225] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:22.378 [2024-11-26 19:26:34.831850] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:22.378 [2024-11-26 19:26:34.832527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.378 [2024-11-26 19:26:34.832565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.378 [2024-11-26 19:26:34.832577] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:22.378 [2024-11-26 19:26:34.832817] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.378 [2024-11-26 19:26:34.833051] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.378 [2024-11-26 19:26:34.833061] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.378 [2024-11-26 19:26:34.833069] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.378 [2024-11-26 19:26:34.833077] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.378 [2024-11-26 19:26:34.838033] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:22.378 [2024-11-26 19:26:34.845713] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.378 [2024-11-26 19:26:34.846284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.378 [2024-11-26 19:26:34.846322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.378 [2024-11-26 19:26:34.846334] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.378 [2024-11-26 19:26:34.846576] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.378 [2024-11-26 19:26:34.846800] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.378 [2024-11-26 19:26:34.846809] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.378 [2024-11-26 19:26:34.846816] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.378 [2024-11-26 19:26:34.846824] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.378 [2024-11-26 19:26:34.859667] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.378 [2024-11-26 19:26:34.860231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.378 [2024-11-26 19:26:34.860268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.378 [2024-11-26 19:26:34.860280] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.378 [2024-11-26 19:26:34.860520] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.378 [2024-11-26 19:26:34.860744] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.378 [2024-11-26 19:26:34.860764] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.378 [2024-11-26 19:26:34.860773] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.378 [2024-11-26 19:26:34.860781] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.378 Malloc0 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:22.378 [2024-11-26 19:26:34.873637] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.378 [2024-11-26 19:26:34.874344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.378 [2024-11-26 19:26:34.874382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.378 [2024-11-26 19:26:34.874394] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.378 [2024-11-26 19:26:34.874634] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.378 [2024-11-26 19:26:34.874858] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.378 [2024-11-26 19:26:34.874876] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.378 [2024-11-26 19:26:34.874884] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.378 [2024-11-26 19:26:34.874892] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:22.378 [2024-11-26 19:26:34.887513] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.378 [2024-11-26 19:26:34.888205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:22.378 [2024-11-26 19:26:34.888243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1d826a0 with addr=10.0.0.2, port=4420 00:30:22.378 [2024-11-26 19:26:34.888254] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d826a0 is same with the state(6) to be set 00:30:22.378 [2024-11-26 19:26:34.888495] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d826a0 (9): Bad file descriptor 00:30:22.378 [2024-11-26 19:26:34.888719] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:30:22.378 [2024-11-26 19:26:34.888727] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:30:22.378 [2024-11-26 19:26:34.888735] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:30:22.378 [2024-11-26 19:26:34.888743] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:22.378 [2024-11-26 19:26:34.900276] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:22.378 [2024-11-26 19:26:34.901391] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:22.378 19:26:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@38 -- # wait 4057001 00:30:22.378 [2024-11-26 19:26:34.969543] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller successful. 00:30:23.887 4861.29 IOPS, 18.99 MiB/s [2024-11-26T18:26:37.449Z] 5638.50 IOPS, 22.03 MiB/s [2024-11-26T18:26:38.393Z] 6246.22 IOPS, 24.40 MiB/s [2024-11-26T18:26:39.333Z] 6771.00 IOPS, 26.45 MiB/s [2024-11-26T18:26:40.720Z] 7172.55 IOPS, 28.02 MiB/s [2024-11-26T18:26:41.662Z] 7501.08 IOPS, 29.30 MiB/s [2024-11-26T18:26:42.600Z] 7786.46 IOPS, 30.42 MiB/s [2024-11-26T18:26:43.541Z] 8021.21 IOPS, 31.33 MiB/s 00:30:30.916 Latency(us) 00:30:30.916 [2024-11-26T18:26:43.541Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:30.916 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:30:30.916 Verification LBA range: start 0x0 length 0x4000 00:30:30.916 Nvme1n1 : 15.01 8234.92 32.17 9798.35 0.00 7072.19 576.85 14417.92 00:30:30.916 [2024-11-26T18:26:43.541Z] =================================================================================================================== 00:30:30.916 [2024-11-26T18:26:43.541Z] Total : 8234.92 32.17 9798.35 0.00 7072.19 576.85 14417.92 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@39 -- # sync 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@42 -- # trap - SIGINT SIGTERM EXIT 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@44 -- # nvmftestfini 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@516 -- # nvmfcleanup 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@121 -- # sync 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@124 -- # set +e 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:30.916 rmmod nvme_tcp 00:30:30.916 rmmod nvme_fabrics 00:30:30.916 rmmod nvme_keyring 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@128 -- # set -e 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@129 -- # return 0 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@517 -- # '[' -n 4058088 ']' 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@518 -- # killprocess 4058088 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@954 -- # '[' -z 4058088 ']' 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@958 -- # kill -0 4058088 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@959 -- # uname 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:30.916 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4058088 00:30:31.177 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:30:31.177 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:30:31.177 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4058088' 00:30:31.178 killing process with pid 4058088 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@973 -- # kill 4058088 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@978 -- # wait 4058088 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@297 -- # iptr 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@791 -- # iptables-save 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@791 -- # iptables-restore 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:31.178 19:26:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:33.720 00:30:33.720 real 0m28.440s 00:30:33.720 user 1m1.499s 00:30:33.720 sys 0m8.153s 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:30:33.720 ************************************ 00:30:33.720 END TEST nvmf_bdevperf 00:30:33.720 ************************************ 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@48 -- # run_test nvmf_target_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:30:33.720 ************************************ 00:30:33.720 START TEST nvmf_target_disconnect 00:30:33.720 ************************************ 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:30:33.720 * Looking for test storage... 00:30:33.720 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:30:33.720 19:26:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1693 -- # lcov --version 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@345 -- # : 1 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:33.720 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # decimal 1 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=1 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 1 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # decimal 2 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=2 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 2 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # return 0 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:30:33.721 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:33.721 --rc genhtml_branch_coverage=1 00:30:33.721 --rc genhtml_function_coverage=1 00:30:33.721 --rc genhtml_legend=1 00:30:33.721 --rc geninfo_all_blocks=1 00:30:33.721 --rc geninfo_unexecuted_blocks=1 00:30:33.721 00:30:33.721 ' 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:30:33.721 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:33.721 --rc genhtml_branch_coverage=1 00:30:33.721 --rc genhtml_function_coverage=1 00:30:33.721 --rc genhtml_legend=1 00:30:33.721 --rc geninfo_all_blocks=1 00:30:33.721 --rc geninfo_unexecuted_blocks=1 00:30:33.721 00:30:33.721 ' 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:30:33.721 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:33.721 --rc genhtml_branch_coverage=1 00:30:33.721 --rc genhtml_function_coverage=1 00:30:33.721 --rc genhtml_legend=1 00:30:33.721 --rc geninfo_all_blocks=1 00:30:33.721 --rc geninfo_unexecuted_blocks=1 00:30:33.721 00:30:33.721 ' 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:30:33.721 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:33.721 --rc genhtml_branch_coverage=1 00:30:33.721 --rc genhtml_function_coverage=1 00:30:33.721 --rc genhtml_legend=1 00:30:33.721 --rc geninfo_all_blocks=1 00:30:33.721 --rc geninfo_unexecuted_blocks=1 00:30:33.721 00:30:33.721 ' 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # uname -s 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@5 -- # export PATH 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.721 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@51 -- # : 0 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:30:33.722 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@11 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@13 -- # MALLOC_BDEV_SIZE=64 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@69 -- # nvmftestinit 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@476 -- # prepare_net_devs 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@438 -- # local -g is_hw=no 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # remove_spdk_ns 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:30:33.722 19:26:46 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # e810=() 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # x722=() 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:41.868 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:30:41.868 Found 0000:31:00.0 (0x8086 - 0x159b) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:30:41.869 Found 0000:31:00.1 (0x8086 - 0x159b) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:30:41.869 Found net devices under 0000:31:00.0: cvl_0_0 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:30:41.869 Found net devices under 0000:31:00.1: cvl_0_1 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # is_hw=yes 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:41.869 19:26:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:41.869 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:41.869 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.623 ms 00:30:41.869 00:30:41.869 --- 10.0.0.2 ping statistics --- 00:30:41.869 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:41.869 rtt min/avg/max/mdev = 0.623/0.623/0.623/0.000 ms 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:41.869 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:41.869 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.325 ms 00:30:41.869 00:30:41.869 --- 10.0.0.1 ping statistics --- 00:30:41.869 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:41.869 rtt min/avg/max/mdev = 0.325/0.325/0.325/0.000 ms 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@450 -- # return 0 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@70 -- # run_test nvmf_target_disconnect_tc1 nvmf_target_disconnect_tc1 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:30:41.869 ************************************ 00:30:41.869 START TEST nvmf_target_disconnect_tc1 00:30:41.869 ************************************ 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1129 -- # nvmf_target_disconnect_tc1 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- host/target_disconnect.sh@32 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@652 -- # local es=0 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect ]] 00:30:41.869 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:30:41.869 [2024-11-26 19:26:54.478534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:41.870 [2024-11-26 19:26:54.478604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1c90d00 with addr=10.0.0.2, port=4420 00:30:41.870 [2024-11-26 19:26:54.478638] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:30:41.870 [2024-11-26 19:26:54.478650] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:30:41.870 [2024-11-26 19:26:54.478659] nvme.c: 951:spdk_nvme_probe_ext: *ERROR*: Create probe context failed 00:30:41.870 spdk_nvme_probe() failed for transport address '10.0.0.2' 00:30:41.870 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect: errors occurred 00:30:41.870 Initializing NVMe Controllers 00:30:41.870 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@655 -- # es=1 00:30:41.870 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:30:41.870 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:30:41.870 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:30:41.870 00:30:41.870 real 0m0.133s 00:30:41.870 user 0m0.060s 00:30:41.870 sys 0m0.072s 00:30:41.870 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:41.870 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@10 -- # set +x 00:30:41.870 ************************************ 00:30:41.870 END TEST nvmf_target_disconnect_tc1 00:30:41.870 ************************************ 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@71 -- # run_test nvmf_target_disconnect_tc2 nvmf_target_disconnect_tc2 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:30:42.131 ************************************ 00:30:42.131 START TEST nvmf_target_disconnect_tc2 00:30:42.131 ************************************ 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1129 -- # nvmf_target_disconnect_tc2 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@37 -- # disconnect_init 10.0.0.2 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # nvmfpid=4064757 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # waitforlisten 4064757 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # '[' -z 4064757 ']' 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:42.131 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:42.131 19:26:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:42.131 [2024-11-26 19:26:54.634875] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:30:42.131 [2024-11-26 19:26:54.634933] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:42.131 [2024-11-26 19:26:54.742548] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:30:42.392 [2024-11-26 19:26:54.794703] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:42.392 [2024-11-26 19:26:54.794777] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:42.392 [2024-11-26 19:26:54.794787] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:42.392 [2024-11-26 19:26:54.794795] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:42.392 [2024-11-26 19:26:54.794801] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:42.392 [2024-11-26 19:26:54.797103] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:30:42.392 [2024-11-26 19:26:54.797264] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:30:42.392 [2024-11-26 19:26:54.797425] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:30:42.392 [2024-11-26 19:26:54.797425] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@868 -- # return 0 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:42.964 Malloc0 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:42.964 [2024-11-26 19:26:55.552503] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:42.964 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:43.226 [2024-11-26 19:26:55.592931] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@42 -- # reconnectpid=4064962 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@44 -- # sleep 2 00:30:43.226 19:26:55 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:30:45.148 19:26:57 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@45 -- # kill -9 4064757 00:30:45.148 19:26:57 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@47 -- # sleep 2 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Write completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 Read completed with error (sct=0, sc=8) 00:30:45.148 starting I/O failed 00:30:45.148 [2024-11-26 19:26:57.628009] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:30:45.148 [2024-11-26 19:26:57.628574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.148 [2024-11-26 19:26:57.628617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.148 qpair failed and we were unable to recover it. 00:30:45.148 [2024-11-26 19:26:57.628842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.148 [2024-11-26 19:26:57.628856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.148 qpair failed and we were unable to recover it. 00:30:45.148 [2024-11-26 19:26:57.629263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.148 [2024-11-26 19:26:57.629298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.148 qpair failed and we were unable to recover it. 00:30:45.148 [2024-11-26 19:26:57.629639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.148 [2024-11-26 19:26:57.629653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.148 qpair failed and we were unable to recover it. 00:30:45.148 [2024-11-26 19:26:57.629895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.148 [2024-11-26 19:26:57.629914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.148 qpair failed and we were unable to recover it. 00:30:45.148 [2024-11-26 19:26:57.630270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.630281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.630558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.630568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.630876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.630888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.631190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.631200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.631539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.631550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.631835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.631846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.632102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.632112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.632289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.632299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.632599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.632608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.632924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.632934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.633258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.633268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.633580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.633591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.633939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.633950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.634264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.634275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.634459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.634469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.634807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.634817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.635227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.635237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.635555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.635565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.635764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.635774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.636090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.636100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.636436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.636447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.636738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.636749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.636952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.636963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.637262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.637272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.637584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.637596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.637925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.637936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.638283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.638294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.638623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.638633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.638843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.638853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.638944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.638954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.639305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.639315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.639604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.639614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.639807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.639819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.640146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.640156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.640506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.640517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.640851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.640865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.641188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.641198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.149 [2024-11-26 19:26:57.641495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.149 [2024-11-26 19:26:57.641505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.149 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.641704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.641715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.641927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.641938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.642263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.642273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.642565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.642575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.642917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.642927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.643262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.643273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.643611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.643621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.643936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.643947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.644291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.644301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.644630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.644640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.644939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.644949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.645269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.645279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.645461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.645471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.645895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.645906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.646121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.646130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.646410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.646419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.646693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.646702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.646883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.646893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.647184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.647195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.647249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.647260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.647462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.647473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.647818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.647828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.648195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.648205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.648530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.648539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.648855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.648874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.649170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.649180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.649476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.649489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.649778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.649788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.650074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.650085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.650385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.650395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.650531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.650541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.650721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.650732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.651021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.651032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.651325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.651336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.651618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.651630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.651794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.651807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.652152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.652165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.150 [2024-11-26 19:26:57.652510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.150 [2024-11-26 19:26:57.652522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.150 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.652906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.652919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.653220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.653233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.653540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.653552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.653883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.653897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.654256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.654269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.654572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.654584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.654925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.654938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.655242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.655254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.655637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.655649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.655820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.655833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.656152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.656166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.656479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.656492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.656678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.656690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.656859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.656894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.657210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.657222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.657550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.657562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.657873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.657887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.658183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.658196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.658488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.658500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.658880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.658894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.659194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.659206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.659528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.659540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.659835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.659847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.660174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.660187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.660531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.660543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.660742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.660755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.661130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.661144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.661479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.661491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.661805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.661820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.662143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.662156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.662477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.662489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.662873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.662886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.663202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.663214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.663527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.663539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.663872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.663885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.664171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.664183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.664468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.664480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.151 [2024-11-26 19:26:57.664794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.151 [2024-11-26 19:26:57.664811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.151 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.665104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.665117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.665311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.665324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.665627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.665641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.666001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.666015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.666333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.666348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.666628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.666642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.666962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.666977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.667167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.667183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.667496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.667510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.667836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.667850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.668149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.668164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.668496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.668511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.668824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.668838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.669197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.669212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.669553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.669567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.669885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.669900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.670219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.670234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.670483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.670499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.670832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.670847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.671146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.671161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.671479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.671494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.671799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.671813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.672133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.672147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.672445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.672460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.672762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.672777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.673109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.673125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.673449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.673464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.673777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.673791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.674106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.674121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.674462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.674477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.674793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.674810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.675176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.675191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.675522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.675536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.675850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.675871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.676277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.676291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.676588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.676603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.676914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.676929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.152 qpair failed and we were unable to recover it. 00:30:45.152 [2024-11-26 19:26:57.677221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.152 [2024-11-26 19:26:57.677236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.677582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.677596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.677807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.677822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.678144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.678159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.678504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.678519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.678766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.678780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.679107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.679123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.679464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.679480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.679658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.679675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.679875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.679891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.680110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.680125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.680436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.680449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.680667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.680682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.681007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.681022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.681221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.681237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.681555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.681569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.681901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.681932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.682245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.682260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.682468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.682482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.682806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.682820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.683193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.683208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.683483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.683497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.683838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.683852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.684220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.684235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.684562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.684576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.684925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.684941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.685254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.685268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.685628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.685643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.685805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.685819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.686146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.686162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.686440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.153 [2024-11-26 19:26:57.686454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.153 qpair failed and we were unable to recover it. 00:30:45.153 [2024-11-26 19:26:57.686768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.686782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.687084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.687099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.687419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.687437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.687751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.687766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.688092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.688107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.688393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.688408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.688740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.688755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.688984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.688998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.689308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.689322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.689537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.689553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.689896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.689912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.690149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.690163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.690522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.690537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.690812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.690826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.691166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.691181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.691468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.691483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.691816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.691831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.692171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.692186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.692480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.692495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.692880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.692897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.693122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.693136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.693453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.693467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.693805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.693819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.694210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.694225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.694528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.694543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.694876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.694891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.695201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.695221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.695526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.695540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.695745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.695759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.696082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.696098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.696431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.696446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.696655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.696669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.696911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.696925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.697206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.697221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.697552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.697566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.697898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.697913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.698199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.698213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.698513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.154 [2024-11-26 19:26:57.698527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.154 qpair failed and we were unable to recover it. 00:30:45.154 [2024-11-26 19:26:57.698837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.698852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.699182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.699198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.699535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.699549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.699759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.699775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.700087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.700106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.700429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.700444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.700768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.700782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.700995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.701010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.701329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.701344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.701675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.701690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.701992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.702008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.702328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.702342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.702671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.702685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.703019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.703034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.703211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.703227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.703571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.703585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.703919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.703935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.704257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.704272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.704596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.704610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.704931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.704946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.705267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.705281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.705434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.705448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.705727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.705741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.706083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.706099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.706425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.706441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.706803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.706817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.707146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.707161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.707494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.707509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.707846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.707860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.708170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.708184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.708557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.708571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.708852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.708879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.709179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.709194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.709506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.709526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.709842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.709857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.710231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.710246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.710571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.710585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.710878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.710893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.155 qpair failed and we were unable to recover it. 00:30:45.155 [2024-11-26 19:26:57.711212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.155 [2024-11-26 19:26:57.711226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.711547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.711570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.711875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.711890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.712190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.712204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.712407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.712422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.712726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.712741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.713077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.713099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.713427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.713442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.713654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.713668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.714029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.714045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.714375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.714389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.714721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.714735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.715110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.715125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.715417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.715433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.715771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.715787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.716092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.716108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.716423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.716437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.716808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.716823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.717135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.717150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.717517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.717531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.717818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.717833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.718137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.718152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.718326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.718341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.718581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.718595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.718893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.718908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.719230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.719245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.719545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.719560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.719948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.719963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.720282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.720296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.720645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.720660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.720983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.720998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.721281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.721295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.721688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.721702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.722035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.722051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.722370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.722384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.722716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.722730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.722928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.722943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.723112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.723127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.156 [2024-11-26 19:26:57.723414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.156 [2024-11-26 19:26:57.723428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.156 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.723782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.723796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.724005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.724020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.724341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.724355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.724696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.724710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.725046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.725061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.725406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.725420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.725794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.725809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.726096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.726114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.726461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.726476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.726800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.726814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.727133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.727148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.727492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.727506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.727836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.727850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.728169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.728184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.728552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.728567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.728735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.728751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.729092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.729107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.729439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.729454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.729745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.729759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.730083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.730098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.730437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.730452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.730821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.730835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.731061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.731076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.731367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.731382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.731722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.731736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.732031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.732046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.732386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.732401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.732736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.732750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.733089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.733104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.733479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.733493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.733796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.733810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.734118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.734133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.734458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.157 [2024-11-26 19:26:57.734472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.157 qpair failed and we were unable to recover it. 00:30:45.157 [2024-11-26 19:26:57.734653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.734669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.735000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.735016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.735338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.735353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.735669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.735683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.735996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.736011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.736318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.736333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.736660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.736674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.737008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.737023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.737363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.737386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.737715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.737729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.738033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.738048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.738363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.738377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.738562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.738578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.738884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.738900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.739095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.739113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.739317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.739331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.739695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.739709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.740027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.740042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.740342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.740356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.740663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.740677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.741001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.741016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.741341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.741355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.741677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.741692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.742086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.742100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.742400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.742414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.742731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.742746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.743084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.743098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.743396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.743410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.743795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.743810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.744129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.744144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.744320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.744336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.744665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.744680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.744994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.745008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.745330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.745344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.745633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.745648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.745990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.746005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.746321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.746336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.746657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.746671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.158 [2024-11-26 19:26:57.747020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.158 [2024-11-26 19:26:57.747036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.158 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.747375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.747390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.747709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.747723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.748054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.748070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.748403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.748417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.748565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.748581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.749001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.749016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.749345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.749360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.749668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.749682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.749972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.749986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.750299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.750314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.750583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.750597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.750983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.750998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.751304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.751319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.751637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.751652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.751946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.751961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.752167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.752182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.752502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.752518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.752842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.752856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.753215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.753230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.753547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.753561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.753790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.753804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.754108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.754123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.754443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.754457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.754820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.754835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.755146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.755161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.755314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.755328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.755616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.755630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.755941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.755956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.756277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.756291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.756597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.756612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.756929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.756943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.757276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.757290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.757600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.757614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.757958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.757973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.758299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.758313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.758533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.758547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.758745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.758760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.159 [2024-11-26 19:26:57.759156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.159 [2024-11-26 19:26:57.759171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.159 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.759467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.759482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.759815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.759829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.760138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.760153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.760493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.760508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.760820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.760839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.761182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.761197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.761535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.761550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.761887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.761902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.762267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.762281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.762609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.762623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.762922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.762937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.763154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.763168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.763359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.763375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.763688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.763702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.764023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.764038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.160 [2024-11-26 19:26:57.764219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.160 [2024-11-26 19:26:57.764235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.160 qpair failed and we were unable to recover it. 00:30:45.435 [2024-11-26 19:26:57.764523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.435 [2024-11-26 19:26:57.764538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.435 qpair failed and we were unable to recover it. 00:30:45.435 [2024-11-26 19:26:57.764852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.435 [2024-11-26 19:26:57.764874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.435 qpair failed and we were unable to recover it. 00:30:45.435 [2024-11-26 19:26:57.765274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.435 [2024-11-26 19:26:57.765289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.435 qpair failed and we were unable to recover it. 00:30:45.435 [2024-11-26 19:26:57.765559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.435 [2024-11-26 19:26:57.765573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.435 qpair failed and we were unable to recover it. 00:30:45.435 [2024-11-26 19:26:57.765826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.435 [2024-11-26 19:26:57.765841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.435 qpair failed and we were unable to recover it. 00:30:45.435 [2024-11-26 19:26:57.766173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.435 [2024-11-26 19:26:57.766189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.435 qpair failed and we were unable to recover it. 00:30:45.435 [2024-11-26 19:26:57.766520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.435 [2024-11-26 19:26:57.766535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.435 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.766867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.766883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.767189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.767205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.767523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.767538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.767731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.767746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.768090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.768105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.768411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.768425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.768757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.768772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.769080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.769094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.769427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.769441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.769751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.769773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.770091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.770106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.770489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.770504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.770838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.770852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.771095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.771109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.771419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.771433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.771721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.771742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.772070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.772085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.772382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.772396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.772714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.772729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.773038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.773053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.773352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.773366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.773688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.773705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.774037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.774052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.774229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.774244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.774589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.774603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.774916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.774931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.775256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.775270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.775456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.775471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.775819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.775834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.776065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.776080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.776411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.776425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.776716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.776731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.777021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.777036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.777388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.777402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.777619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.777633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.777845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.777860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.778202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.778216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.436 qpair failed and we were unable to recover it. 00:30:45.436 [2024-11-26 19:26:57.778552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.436 [2024-11-26 19:26:57.778566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.778870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.778885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.779219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.779233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.779568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.779583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.779880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.779895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.780211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.780225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.780568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.780582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.781008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.781023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.781323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.781337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.781628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.781642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.782019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.782034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.782374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.782388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.782612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.782627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.782946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.782961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.783284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.783306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.783594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.783608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.783934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.783949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.784158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.784173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.784515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.784530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.784855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.784876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.785190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.785204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.785531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.785545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.785881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.785896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.786222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.786237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.786525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.786542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.786827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.786842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.787178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.787193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.787579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.787593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.787933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.787948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.788284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.788298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.788625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.788639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.788871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.788886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.437 qpair failed and we were unable to recover it. 00:30:45.437 [2024-11-26 19:26:57.789093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.437 [2024-11-26 19:26:57.789108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.789430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.789445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.789787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.789801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.790117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.790133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.790376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.790391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.790674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.790688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.791021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.791036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.791347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.791361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.791704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.791719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.791916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.791932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.792271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.792285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.792664] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.792678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.793098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.793113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.793324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.793339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.793673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.793688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.794056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.794071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.794400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.794414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.794793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.794807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.795093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.795107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.795427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.795441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.795618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.795633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.795947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.795962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.796321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.796335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.796671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.796685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.796976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.796990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.797367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.797381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.797670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.797684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.798109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.798125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.798440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.798454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.798796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.798810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.799118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.799132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.799469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.799484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.799770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.799788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.800010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.800026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.800335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.800349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.438 qpair failed and we were unable to recover it. 00:30:45.438 [2024-11-26 19:26:57.800649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.438 [2024-11-26 19:26:57.800663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.800998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.801014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.801255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.801269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.801578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.801592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.801923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.801938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.802256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.802270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.802608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.802622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.802927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.802943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.803249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.803263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.803644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.803658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.803995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.804010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.804338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.804353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.804650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.804664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.804982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.804997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.805294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.805309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.805625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.805640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.805937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.805952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.806329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.806344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.806661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.806675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.807010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.807026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.807326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.807340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.807651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.807665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 [2024-11-26 19:26:57.807766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.439 [2024-11-26 19:26:57.807779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.439 qpair failed and we were unable to recover it. 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Write completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 [2024-11-26 19:26:57.808538] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.439 starting I/O failed 00:30:45.439 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Write completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 Read completed with error (sct=0, sc=8) 00:30:45.440 starting I/O failed 00:30:45.440 [2024-11-26 19:26:57.808784] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:30:45.440 [2024-11-26 19:26:57.809059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.809079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.809410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.809425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.809767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.809781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.810178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.810193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.810500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.810514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.810858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.810879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.811199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.811214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.811517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.811532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.811856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.811875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.812192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.812206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.812528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.812542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.812866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.812881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.813201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.813215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.813537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.813552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.440 [2024-11-26 19:26:57.813733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.440 [2024-11-26 19:26:57.813748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.440 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.814068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.814083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.814418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.814432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.814715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.814731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.815056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.815071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.815418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.815433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.815767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.815781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.816003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.816018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.816363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.816377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.816690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.816705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.817041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.817055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.817391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.817405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.817752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.817766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.818098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.818117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.818455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.818469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.818685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.818699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.818873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.818889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.819227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.819241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.819552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.819566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.819847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.819867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.820060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.820074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.820359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.820373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.820711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.820726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.821042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.821057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.821402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.821417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.821754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.821769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.822100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.822115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.822315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.822330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.822561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.822576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.822872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.822887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.823195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.441 [2024-11-26 19:26:57.823209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.441 qpair failed and we were unable to recover it. 00:30:45.441 [2024-11-26 19:26:57.823395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.823411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.823734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.823748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.824081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.824096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.824392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.824406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.824704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.824718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.825008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.825023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.825339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.825353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.825573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.825587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.825912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.825928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.826254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.826270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.826568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.826582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.826903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.826918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.827144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.827159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.827458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.827473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.827803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.827817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.827998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.828013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.828349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.828364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.828690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.828705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.829011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.829026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.829311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.829325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.829656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.829670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.829949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.829965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.830294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.830312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.830631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.830645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.830944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.830959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.831179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.831193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.831525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.831540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.831851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.831870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.832169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.832183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.832504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.832518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.832837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.832851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.833165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.833180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.833402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.442 [2024-11-26 19:26:57.833416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.442 qpair failed and we were unable to recover it. 00:30:45.442 [2024-11-26 19:26:57.833705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.833724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.833937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.833953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.834277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.834291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.834593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.834608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.834939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.834955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.835345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.835360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.835707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.835722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.836048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.836063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.836379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.836393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.836460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.836475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.836762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.836777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.836999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.837014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.837329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.837344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.837674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.837689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.837886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.837902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.838104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.838119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.838439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.838454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.838747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.838761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.838963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.838978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.839297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.839311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.839626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.839639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.839958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.839973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.840298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.840313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.840651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.840666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.841033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.841048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.841335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.841350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.841677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.841691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.842004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.842019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.842349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.842364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.842702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.842720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.843031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.843046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.843262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.843276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.843606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.843620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.843839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.843853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.844197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.443 [2024-11-26 19:26:57.844211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.443 qpair failed and we were unable to recover it. 00:30:45.443 [2024-11-26 19:26:57.844499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.844513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.844800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.844814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.845191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.845205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.845541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.845555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.845885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.845901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.846192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.846207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.846496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.846510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.846837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.846851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.847228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.847243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.847581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.847595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.847829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.847842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.848180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.848195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.848505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.848519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.848838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.848852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.849177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.849191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.849534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.849548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.849790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.849804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.850120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.850136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.850470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.850484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.850809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.850823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.851110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.851125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.851441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.851456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.851768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.851783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.852090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.852105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.852444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.852459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.852639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.852655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.852954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.852969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.853265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.853279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.853518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.853532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.853880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.853896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.854255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.854270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.854447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.854462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.854789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.854804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.855185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.855200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.855517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.855535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.855901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.855917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.856269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.856282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.856505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.444 [2024-11-26 19:26:57.856520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.444 qpair failed and we were unable to recover it. 00:30:45.444 [2024-11-26 19:26:57.856840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.856854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.857168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.857183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.857392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.857407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.857716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.857730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.858044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.858059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.858385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.858399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.858733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.858747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.859078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.859093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.859431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.859445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.859792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.859806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.860074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.860089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.860427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.860441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.860816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.860830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.861032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.861048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.861266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.861280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.861607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.861622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.861961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.861977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.862303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.862318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.862533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.862547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.862875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.862889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.863223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.863237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.863613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.863627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.863957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.863971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.864181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.864196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.864505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.864519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.864866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.864882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.865230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.865244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.865527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.865547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.865879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.865894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.866211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.866224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.866523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.866537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.866879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.866894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.445 [2024-11-26 19:26:57.867122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.445 [2024-11-26 19:26:57.867136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.445 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.867447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.867461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.867800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.867814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.868168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.868183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.868520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.868538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.868856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.868875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.869206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.869220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.869509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.869524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.869855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.869880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.870187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.870202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.870511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.870525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.870823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.870837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.871167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.871182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.871400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.871415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.871568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.871582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.871918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.871933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.872255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.872270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.872602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.872616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.872898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.872913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.873140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.873155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.873486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.873500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.873817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.873832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.874163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.874178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.874511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.874526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.874858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.874878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.875069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.875084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.875288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.875302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.875625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.875639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.876020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.876036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.876345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.876359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.876676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.876697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.877053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.877069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.877399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.877413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.877738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.877752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.878071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.878086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.878214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.878227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.878441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.878456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.878801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.878815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.446 qpair failed and we were unable to recover it. 00:30:45.446 [2024-11-26 19:26:57.879135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.446 [2024-11-26 19:26:57.879158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.879488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.879503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.879790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.879805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.879993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.880010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.880348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.880362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.880700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.880716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.881014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.881033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.881367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.881382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.881756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.881772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.882076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.882091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.882410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.882424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.882644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.882659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.882988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.883003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.883327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.883342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.883651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.883665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.883973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.883988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.884335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.884350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.884666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.884680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.884997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.885012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.885320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.885335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.885672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.885686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.886001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.886017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.886320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.886335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.886696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.886711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.887137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.887153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.887527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.887542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.887878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.887894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.888223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.888237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.888567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.888584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.888907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.888921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.889265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.889286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.889611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.889626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.889919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.889938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.890138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.890155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.890443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.890457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.890793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.890808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.891122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.891138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.891467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.891481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.447 [2024-11-26 19:26:57.891874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.447 [2024-11-26 19:26:57.891890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.447 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.892212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.892226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.892611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.892625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.892838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.892853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.893159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.893173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.893500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.893515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.893835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.893850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.894073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.894088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.894405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.894423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.894761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.894776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.895084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.895100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.895423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.895437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.895730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.895744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.896085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.896101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.896434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.896450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.896636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.896652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.896947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.896963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.897302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.897317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.897649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.897664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.898039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.898055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.898376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.898391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.898699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.898714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.898942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.898958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.899180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.899195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.899520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.899534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.899852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.899872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.900269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.900284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.900605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.900619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.901043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.901059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.901345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.901360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.901659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.901674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.901995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.902010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.902319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.902333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.902650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.902664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.902869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.902884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.903180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.903195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.903505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.903519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.903814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.903828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.448 qpair failed and we were unable to recover it. 00:30:45.448 [2024-11-26 19:26:57.904155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.448 [2024-11-26 19:26:57.904171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.904492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.904507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.904848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.904869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.905199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.905214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.905542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.905557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.905791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.905807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.906126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.906142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.906332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.906347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.906672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.906688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.907020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.907036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.907340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.907357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.907665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.907680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.907858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.907882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.908170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.908186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.908517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.908532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.908816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.908831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.909157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.909173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.909487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.909501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.909813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.909828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.910001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.910018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.910371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.910386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.910718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.910732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.911067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.911083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.911389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.911404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.911555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.911572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.911918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.911934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.912143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.912158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.912487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.912502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.912843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.912858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.913181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.913195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.913489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.913504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.913801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.913815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.914205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.914221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.914550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.914566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.914901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.914917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.915303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.915318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.915614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.915636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.916003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.449 [2024-11-26 19:26:57.916019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.449 qpair failed and we were unable to recover it. 00:30:45.449 [2024-11-26 19:26:57.916202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.916218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.916558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.916573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.916883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.916898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.917095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.917111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.917436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.917451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.917749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.917763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.918041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.918056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.918393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.918408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.918618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.918634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.918969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.918984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.919338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.919353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.919570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.919584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.919919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.919937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.920256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.920271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.920606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.920621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.920934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.920950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.921140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.921155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.921475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.921489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.921838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.921853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.922255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.922271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.922593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.922608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.922953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.922969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.923277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.923291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.923602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.923616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.923946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.923961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.924279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.924294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.924600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.924615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.924977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.924993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.925378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.925392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.925689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.925704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.926042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.926058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.926383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.926398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.926714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.926729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.450 [2024-11-26 19:26:57.927065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.450 [2024-11-26 19:26:57.927080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.450 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.927427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.927442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.927765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.927780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.928128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.928143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.928447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.928461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.928629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.928646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.928991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.929008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.929362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.929377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.929714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.929729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.929913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.929929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.930228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.930243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.930461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.930475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.930792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.930807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.931123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.931139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.931403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.931417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.931729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.931744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.932085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.932101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.932294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.932312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.932586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.932601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.932930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.932948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.933162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.933178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.933548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.933563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.933883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.933899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.934120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.934136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.934434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.934449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.934694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.934709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.935030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.935047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.935401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.935415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.935732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.935747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.935964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.935979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.936307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.936321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.936621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.936635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.936807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.936823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.937175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.937190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.937508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.937523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.937822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.937837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.938045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.451 [2024-11-26 19:26:57.938060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.451 qpair failed and we were unable to recover it. 00:30:45.451 [2024-11-26 19:26:57.938264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.938280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.938608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.938622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.938919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.938935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.939265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.939280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.939555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.939569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.939763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.939780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.940100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.940115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.940476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.940491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.940777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.940791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.941115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.941131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.941507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.941522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.941915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.941932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.942113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.942127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.942478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.942494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.942820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.942835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.943153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.943176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.943533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.943548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.943874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.943889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.944174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.944188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.944487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.944501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.944849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.944871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.945170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.945185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.945429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.945446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.945796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.945810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.946012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.946030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.946226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.946241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.946632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.946646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.946934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.946949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.947285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.947300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.947604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.947619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.947985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.948001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.948324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.948338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.452 [2024-11-26 19:26:57.948528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.452 [2024-11-26 19:26:57.948543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.452 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.948829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.948844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.949163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.949179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.949477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.949493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.949779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.949794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.950009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.950025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.950362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.950377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.950555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.950572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.950790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.950805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.951133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.951149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.951454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.951469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.951801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.951816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.952124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.952140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.952461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.952484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.952685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.952700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.953005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.953021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.953244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.953258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.953591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.953606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.953914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.953930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.954160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.954175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.954472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.954487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.954811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.954825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.955126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.955142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.955529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.955544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.955884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.955900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.955967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.955982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.956284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.956299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.956637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.956652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.957014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.957030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.957335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.957350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.957671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.957689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.958077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.958094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.958296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.453 [2024-11-26 19:26:57.958311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.453 qpair failed and we were unable to recover it. 00:30:45.453 [2024-11-26 19:26:57.958462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.958478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.958831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.958846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.959225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.959240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.959551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.959565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.959945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.959960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.960262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.960277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.960588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.960602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.960927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.960943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.961295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.961310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.961624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.961638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.961977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.961992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.962356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.962371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.962715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.962730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.963040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.963056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.963354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.963368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.963687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.963702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.963898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.963914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.964256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.964271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.964559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.964573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.964878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.964893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.965212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.965225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.965555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.965570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.965881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.965896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.966229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.966244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.966581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.966596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.966964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.966979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.967248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.967263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.967574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.967589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.967934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.967956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.968150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.968166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.968538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.968553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.968866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.968882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.969208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.969223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.969538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.969554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.969876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.969891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.454 qpair failed and we were unable to recover it. 00:30:45.454 [2024-11-26 19:26:57.970277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.454 [2024-11-26 19:26:57.970292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.970636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.970650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.970983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.971002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.971350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.971365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.971674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.971688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.972023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.972038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.972360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.972374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.972579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.972593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.972912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.972927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.973280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.973295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.973619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.973633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.973931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.973946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.974276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.974290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.974626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.974641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.974974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.974990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.975310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.975332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.975590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.975605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.975783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.975798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.976133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.976148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.976453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.976467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.976780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.976795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.977096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.977112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.977299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.977314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.977529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.977544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.977890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.977905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.978227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.978242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.978561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.978576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.978915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.978931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.979292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.979307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.979649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.979666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.979984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.979999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.980383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.980397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.455 [2024-11-26 19:26:57.980705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.455 [2024-11-26 19:26:57.980719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.455 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.981098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.981113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.981399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.981419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.981732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.981747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.982035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.982050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.982237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.982253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.982474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.982488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.982886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.982903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.983211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.983227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.986572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.986607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.986848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.986887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.987255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.987272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.987651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.987667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.987976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.987992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.988175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.988191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.988507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.988522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.988741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.988756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.988956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.988971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.989375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.989390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.989681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.989697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.990017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.990032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.990239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.990255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.990541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.990555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.990859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.990888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.991109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.991125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.991424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.991440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.991761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.991776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.992105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.992122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.992410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.992425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.992766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.992780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.993102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.993117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.993456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.993471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.993760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.993781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.994108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.994123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.994457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.994472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.994783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.456 [2024-11-26 19:26:57.994797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.456 qpair failed and we were unable to recover it. 00:30:45.456 [2024-11-26 19:26:57.995108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.995123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.995322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.995341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.995690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.995704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.995996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.996011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.996332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.996347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.996651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.996667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.997031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.997046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.997431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.997446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.997785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.997799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.998115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.998131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.998449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.998463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.998766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.998780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.999066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.999084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.999405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.999420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:57.999795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:57.999810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.000059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.000075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.000467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.000482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.000779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.000795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.000967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.000984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.001175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.001192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.001490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.001505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.001808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.001823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.002166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.002182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.002519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.002534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.002844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.002860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.003080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.003095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.003423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.003438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.003774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.003790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.004118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.004134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.004329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.004346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.004659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.004674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.005011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.005026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.005324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.005338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.005653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.005668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.005991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.006006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.006346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.006361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.006661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.457 [2024-11-26 19:26:58.006676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.457 qpair failed and we were unable to recover it. 00:30:45.457 [2024-11-26 19:26:58.007027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.007042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.007449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.007463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.007740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.007754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.008077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.008093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.008386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.008404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.008615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.008631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.008968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.008989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.009292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.009306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.009631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.009645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.009934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.009949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.010275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.010296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.010604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.010618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.010984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.011000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.011321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.011336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.011540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.011554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.011873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.011888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.012227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.012241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.012600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.012614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.012931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.012946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.013306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.013320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.013616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.013631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.013944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.013959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.014289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.014303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.014638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.014653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.014792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.014808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.015162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.015177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.015506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.015521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.015869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.015884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.016208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.016222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.016545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.016559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.016874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.016889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.017230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.017244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.017537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.017552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.017734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.017749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.017925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.017942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.018281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.018296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.018629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.018644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.018970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.458 [2024-11-26 19:26:58.018985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.458 qpair failed and we were unable to recover it. 00:30:45.458 [2024-11-26 19:26:58.019291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.019311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.019694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.019708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.019909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.019926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.020209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.020224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.020545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.020560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.020778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.020793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.020982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.021002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.021333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.021347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.021639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.021653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.021976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.021992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.022203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.022217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.022492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.022507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.022884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.022900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.023225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.023240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.023381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.023397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.023740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.023755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.024078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.024093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.024469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.024484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.024834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.024848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.025152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.025167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.025557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.025572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.025883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.025898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.026245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.026260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.026609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.026625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.026967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.026982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.027302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.027324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.027653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.027667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.027853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.027875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.028113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.028128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.028437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.028452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.028743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.028758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.029083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.029098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.029385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.029407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.029728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.029743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.030059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.030074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.030395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.030410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.030802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.030817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.031115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.459 [2024-11-26 19:26:58.031130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.459 qpair failed and we were unable to recover it. 00:30:45.459 [2024-11-26 19:26:58.031468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.031483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.031678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.031695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.031889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.031906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.032219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.032234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.032597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.032612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.032954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.032969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.033189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.033203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.033522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.033537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.033883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.033901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.034238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.034252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.034607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.034620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.034838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.034854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.035274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.035288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.035579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.035593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.035928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.035942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.036279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.036293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.036597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.036610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.036775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.036788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.037152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.037167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.037484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.037499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.037824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.037839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.038178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.038193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.038545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.038559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.038875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.038890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.039217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.039233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.039568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.039584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.039896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.039912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.040234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.040249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.040580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.040596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.040998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.041015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.041223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.041239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.041545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.041560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.041901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.460 [2024-11-26 19:26:58.041917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.460 qpair failed and we were unable to recover it. 00:30:45.460 [2024-11-26 19:26:58.042246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.042262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.042598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.042614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.042921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.042938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.043322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.043338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.043665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.043681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.043989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.044004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.044211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.044228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.044571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.044586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.044798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.044814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.461 [2024-11-26 19:26:58.045038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.461 [2024-11-26 19:26:58.045054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.461 qpair failed and we were unable to recover it. 00:30:45.763 [2024-11-26 19:26:58.045378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.763 [2024-11-26 19:26:58.045395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.763 qpair failed and we were unable to recover it. 00:30:45.763 [2024-11-26 19:26:58.045579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.045598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.045870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.045886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.046229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.046244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.046568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.046583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.046922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.046947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.047278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.047294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.047631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.047647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.047971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.047987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.048288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.048304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.048629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.048645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.048942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.048958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.049274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.049290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.049613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.049629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.049958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.049975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.050265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.050280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.050617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.050633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.050812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.050828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.051152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.051170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.051360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.051377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.051709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.051725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.051922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.051939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.052126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.052141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.052434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.052450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.052660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.052676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.053004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.053020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.053331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.053348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.053662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.053678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.054022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.054039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.054341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.054357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.054652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.054668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.054999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.055015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.055365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.055381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.055681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.055697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.055978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.055993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.056321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.056336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.056673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.056687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.764 [2024-11-26 19:26:58.057023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.764 [2024-11-26 19:26:58.057038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.764 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.057377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.057393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.057774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.057790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.058001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.058017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.058336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.058350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.058648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.058662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.058916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.058931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.059152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.059173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.059523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.059542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.059828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.059844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.060189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.060206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.060580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.060595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.060929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.060946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.061279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.061294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.061610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.061626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.061911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.061928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.062272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.062287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.062603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.062617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.062962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.062978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.063293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.063307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.063640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.063655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.063938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.063954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.064264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.064280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.064502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.064517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.064858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.064880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.065205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.065220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.065530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.065545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.065924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.065941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.066290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.066305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.066597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.066612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.066952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.066969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.067284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.067298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.067635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.067650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.067985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.068001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.068334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.068349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.068684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.068700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.069091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.069108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.069435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.069450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.069784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.765 [2024-11-26 19:26:58.069798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.765 qpair failed and we were unable to recover it. 00:30:45.765 [2024-11-26 19:26:58.069928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.069943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.070298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.070312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.070604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.070619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.070819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.070834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.071168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.071184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.071505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.071521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.071817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.071831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.072151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.072167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.072460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.072474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.072794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.072812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.073181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.073198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.073533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.073548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.073841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.073857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.074193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.074207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.074500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.074515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.074871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.074887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.075085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.075101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.075450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.075465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.075797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.075811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.075998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.076015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.076352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.076367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.076669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.076685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.077023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.077040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.077335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.077351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.077643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.077658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.078006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.078022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.078381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.078395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.078728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.078743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.079051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.079066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.079383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.079397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.079713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.079727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.080036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.080051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.080396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.080411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.080725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.080740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.081111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.081127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.081409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.081423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.081764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.081779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.082096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.766 [2024-11-26 19:26:58.082111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.766 qpair failed and we were unable to recover it. 00:30:45.766 [2024-11-26 19:26:58.082451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.082465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.082664] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.082680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.082996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.083012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.083388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.083403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.083736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.083751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.083943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.083958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.084337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.084352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.084677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.084691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.085024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.085039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.085362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.085376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.085718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.085733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.086082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.086100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.086331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.086347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.086671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.086686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.086996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.087011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.087343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.087357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.087667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.087682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.088001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.088017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.088403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.088417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.088709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.088730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.089044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.089059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.089348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.089369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.089677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.089691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.089983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.089999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.090315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.090330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.090650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.090665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.091007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.091023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.091331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.091346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.091670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.091685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.092016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.092031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.092330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.092345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.092697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.092711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.092996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.093012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.093358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.093372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.093670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.093684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.093894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.093910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.094250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.094264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.094492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.094506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.767 qpair failed and we were unable to recover it. 00:30:45.767 [2024-11-26 19:26:58.094866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.767 [2024-11-26 19:26:58.094882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.095259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.095273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.095588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.095602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.095929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.095944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.096154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.096168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.096511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.096526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.096856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.096878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.097211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.097225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.097602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.097616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.097926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.097944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.098292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.098307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.098523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.098538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.098721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.098736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.099058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.099077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.099377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.099392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.099706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.099721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.100017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.100032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.100350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.100365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.100701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.100717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.101037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.101052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.101345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.101361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.101675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.101689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.102026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.102041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.102343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.102358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.102676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.102691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.103030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.103046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.103391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.103406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.103729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.103743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.104150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.104165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.104474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.104488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.104812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.104827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.105215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.105230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.105535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.768 [2024-11-26 19:26:58.105549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.768 qpair failed and we were unable to recover it. 00:30:45.768 [2024-11-26 19:26:58.105874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.105889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.106217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.106232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.106554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.106568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.106883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.106898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.107105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.107119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.107451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.107465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.107841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.107856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.108173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.108188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.108539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.108554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.108847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.108869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.109096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.109111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.109426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.109440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.109619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.109635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.109973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.109989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.110316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.110337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.110645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.110660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.111011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.111026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.111367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.111381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.111705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.111728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.112054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.112069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.112447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.112468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.112792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.112806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.113145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.113160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.113338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.113353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.113693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.113707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.113920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.113935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.114208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.114223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.114552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.114566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.114761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.114777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.115121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.115136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.115428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.115442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.115744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.115758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.116169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.116184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.116527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.116541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.116857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.116877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.117206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.117222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.117561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.117575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.117759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.117775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.769 qpair failed and we were unable to recover it. 00:30:45.769 [2024-11-26 19:26:58.117971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.769 [2024-11-26 19:26:58.117987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.118268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.118283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.118576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.118599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.118908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.118924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.119260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.119274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.119617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.119631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.119936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.119951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.120264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.120279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.120596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.120610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.120800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.120816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.121156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.121171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.121486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.121500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.121828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.121842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.122256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.122271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.122633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.122647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.122993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.123009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.123339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.123354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.123688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.123703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.124028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.124043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.124383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.124397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.124756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.124771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.125073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.125088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.125471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.125489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.125835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.125849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.126089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.126103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.126422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.126436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.126669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.126685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.126988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.127003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.127327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.127341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.127677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.127692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.128022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.128037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.128232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.128248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.128584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.128600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.128845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.128860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.129257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.129272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.129568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.129583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.129922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.129939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.130257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.130273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.770 [2024-11-26 19:26:58.130588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.770 [2024-11-26 19:26:58.130604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.770 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.130787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.130804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.131139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.131156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.131470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.131486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.131836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.131851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.132221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.132237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.132448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.132463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.132771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.132786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.133148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.133165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.133505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.133521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.133821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.133835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.134222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.134239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.134565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.134580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.134955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.134972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.135254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.135268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.135462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.135479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.135802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.135817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.136196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.136212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.136538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.136554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.136833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.136847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.137203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.137219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.137409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.137425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.137709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.137725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.138055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.138071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.138361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.138380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.138592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.138609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.138941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.138957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.139273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.139289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.139609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.139625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.139907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.139923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.140256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.140272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.140506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.140521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.140740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.140754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.141080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.141096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.141314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.141329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.141688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.141702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.141985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.771 [2024-11-26 19:26:58.142001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.771 qpair failed and we were unable to recover it. 00:30:45.771 [2024-11-26 19:26:58.142323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.142338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.142658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.142674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.143017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.143033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.143215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.143231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.143560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.143575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.143797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.143812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.144200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.144216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.144533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.144549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.144889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.144905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.145213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.145227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.145522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.145536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.145755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.145769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.146075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.146091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.146460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.146475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.146870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.146890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.147199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.147214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.147563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.147578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.147910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.147926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.148282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.148298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.148613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.148627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.148964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.148979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.149299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.149315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.149645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.149662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.149873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.149891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.150252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.150267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.150596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.150612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.150916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.150932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.151280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.151295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.151625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.151641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.151969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.151985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.152388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.152402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.152725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.152740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.153079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.153096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.153384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.153398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.153611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.153625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.153957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.772 [2024-11-26 19:26:58.153972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.772 qpair failed and we were unable to recover it. 00:30:45.772 [2024-11-26 19:26:58.154295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.154311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.154621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.154637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.155031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.155047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.155340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.155355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.155691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.155708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.155921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.155938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.156284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.156299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.156510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.156526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.156875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.156891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.157230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.157244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.157633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.157648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.157940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.157956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.158246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.158261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.158591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.158607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.158816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.158833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.159179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.159194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.159526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.159542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.159883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.159900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.160197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.160214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.160567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.160583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.161000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.161016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.161358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.161376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.161589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.161605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.161920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.161937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.162246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.162263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.162589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.162605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.162932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.162950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.163286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.163303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.163642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.163659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.164029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.164046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.773 qpair failed and we were unable to recover it. 00:30:45.773 [2024-11-26 19:26:58.164391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.773 [2024-11-26 19:26:58.164408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.164728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.164745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.164921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.164940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.165270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.165286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.165618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.165634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.165971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.165988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.166321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.166337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.166666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.166684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.166997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.167014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.167332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.167351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.167661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.167679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.168001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.168018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.168356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.168372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.168687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.168705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.169046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.169065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.169403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.169422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.169757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.169775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.170093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.170111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.170323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.170342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.170660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.170678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.171056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.171075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.171398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.171416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.171720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.171739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.171953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.171972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.172207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.172226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.172535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.172553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.172892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.172911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.173259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.173276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.173578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.173598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.173923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.173942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.174259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.174277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.174609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.174627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.174975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.174994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.175323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.175342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.175653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.175672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.176007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.176027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.774 [2024-11-26 19:26:58.176363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.774 [2024-11-26 19:26:58.176381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.774 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.176697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.176717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.177025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.177045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.177369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.177387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.177705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.177723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.178034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.178053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.178366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.178385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.178705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.178723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.179045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.179065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.179380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.179398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.179703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.179721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.180040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.180059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.180408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.180427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.180728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.180746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.181078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.181097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.181411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.181429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.181744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.181762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.182166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.182184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.182514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.182531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.182873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.182893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.183211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.183230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.183540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.183558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.183741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.183761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.184091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.184110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.184428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.184447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.184789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.184807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.185129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.185148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.185477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.185495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.185807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.185825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.186117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.186135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.186444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.186463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.186796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.186815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.186993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.187015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.187372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.187389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.775 [2024-11-26 19:26:58.187701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.775 [2024-11-26 19:26:58.187717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.775 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.188049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.188066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.188379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.188396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.188729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.188746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.189048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.189064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.189293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.189310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.189630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.189647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.189979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.189997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.190344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.190362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.190683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.190700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.191039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.191056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.191364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.191381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.191686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.191702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.192037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.192055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.192360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.192376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.192712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.192729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.193067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.193084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.193409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.193426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.193732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.193748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.193954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.193972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.194357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.194375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.194700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.194716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.195068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.195086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.195399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.195418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.195728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.195745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.196050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.196067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.196420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.196437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.196617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.196633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.196973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.196991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.197320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.197339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.197647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.197664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.197892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.197909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.198223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.198240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.198571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.198588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.198783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.198802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.199098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.199115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.776 [2024-11-26 19:26:58.199419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.776 [2024-11-26 19:26:58.199437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.776 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.199781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.199798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.200113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.200136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.200445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.200462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.200804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.200822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.201143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.201160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.201414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.201431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.201775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.201793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.202003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.202020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.202355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.202371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.202704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.202722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.203054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.203071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.203409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.203426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.203758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.203774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.204105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.204122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.204425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.204441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.204659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.204676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.205003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.205021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.205334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.205353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.205683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.205700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.206027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.206044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.206386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.206403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.206740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.206759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.207078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.207096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.207436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.207454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.207785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.207802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.208129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.208147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.208448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.208465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.208810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.208829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.209134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.209152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.209464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.209481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.209822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.209839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.209998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.210015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.210339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.210356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.210663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.210682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.211023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.777 [2024-11-26 19:26:58.211041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.777 qpair failed and we were unable to recover it. 00:30:45.777 [2024-11-26 19:26:58.211361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.211379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.211528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.211545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.211840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.211856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.212163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.212181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.212500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.212516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.212832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.212850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.213167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.213188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.213499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.213517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.213853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.213877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.214106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.214122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.214428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.214445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.214788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.214805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.215086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.215103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.215439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.215456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.215798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.215814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.216128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.216146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.216475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.216492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.216831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.216848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.217154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.217171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.217488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.217506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.217843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.217867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.218189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.218206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.218501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.218519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.218874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.218893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.219193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.219210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.219425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.219442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.219737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.219753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.220088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.220106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.220417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.220434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.220769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.220786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.221087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.221105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.221455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.221473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.221822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.221840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.222144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.222163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.222383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.222398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.778 qpair failed and we were unable to recover it. 00:30:45.778 [2024-11-26 19:26:58.222721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.778 [2024-11-26 19:26:58.222738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.223038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.223056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.223391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.223408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.223714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.223730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.224081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.224099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.224433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.224451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.224788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.224804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.225121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.225138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.225503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.225520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.225826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.225842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.226170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.226188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.226536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.226557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.226893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.226911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.227184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.227200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.227523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.227541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.227727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.227745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.228077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.228095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.228437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.228453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.228783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.228800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.229095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.229112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.229464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.229480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.229775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.229792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.230129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.230146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.230485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.230503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.230897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.230914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.231239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.231256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.231587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.231604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.231781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.231798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.232110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.232128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.232462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.232479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.232843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.232860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.233171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.233187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.233376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.233394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.233723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.233739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.234140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.234158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.234488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.234506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.234731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.779 [2024-11-26 19:26:58.234749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.779 qpair failed and we were unable to recover it. 00:30:45.779 [2024-11-26 19:26:58.235105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.235124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.235463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.235480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.235818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.235836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.236193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.236211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.236534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.236550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.236886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.236904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.237236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.237254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.237576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.237593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.237931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.237948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.238316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.238334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.238652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.238671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.238990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.239008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.239344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.239362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.239695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.239713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.240085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.240107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.240427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.240444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.240794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.240813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.241175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.241193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.241516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.241534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.241622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.241638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.241955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.241973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.242310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.242328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.242622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.242639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.242994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.243013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.243325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.243342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.243721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.243738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.244037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.244055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.244360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.244377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.244690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.244707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.245037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.245055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.245359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.245376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.245714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.245731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.246072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.246091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.246422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.246441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.246800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.246817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.247158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.247176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.247480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.247497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.247874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.780 [2024-11-26 19:26:58.247892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.780 qpair failed and we were unable to recover it. 00:30:45.780 [2024-11-26 19:26:58.248228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.248244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.248572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.248589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.248909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.248926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.249279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.249296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.249647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.249666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.249992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.250009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.250344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.250362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.250695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.250713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.251014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.251031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.251373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.251389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.251698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.251715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.251904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.251923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.252261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.252277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.252609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.252626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.252945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.252963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.253307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.253324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.253635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.253656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.253985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.254003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.254208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.254225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.254557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.254574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.254882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.254900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.255237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.255254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.255594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.255612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.255946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.255964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.256337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.256355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.256676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.256693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.257018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.257036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.257378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.257395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.257717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.257735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.257935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.257953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.258330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.258349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.258681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.258698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.259037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.259055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.259240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.259257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.259522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.259538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.259874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.259892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.260239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.260258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.260602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.781 [2024-11-26 19:26:58.260619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.781 qpair failed and we were unable to recover it. 00:30:45.781 [2024-11-26 19:26:58.260960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.260979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.261337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.261354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.261655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.261671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.261976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.261993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.262310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.262328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.262660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.262678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.262984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.263000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.263215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.263236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.263566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.263584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.263807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.263823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.264189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.264207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.264393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.264411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.264732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.264750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.264962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.264980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.265321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.265338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.265669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.265687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.265879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.265899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.266187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.266204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.266542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.266563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.266873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.266890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.267096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.267115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.267317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.267335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.267637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.267657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.267964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.267983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.268372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.268389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.268720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.268737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.269076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.269094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.269303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.269320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.269646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.269664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.270003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.270022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.270334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.270351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.270665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.270683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.271001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.271019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.782 [2024-11-26 19:26:58.271343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.782 [2024-11-26 19:26:58.271362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.782 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.271689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.271706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.272024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.272042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.272388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.272405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.272585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.272602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.272935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.272954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.273179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.273196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.273527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.273546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.273884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.273904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.274226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.274244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.274543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.274562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.274874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.274892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.275220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.275240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.275544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.275561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.275884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.275903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.276245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.276262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.276593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.276611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.276803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.276821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.277158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.277177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.277494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.277511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.277855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.277891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.278240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.278256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.278601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.278618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.278931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.278948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.279178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.279195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.279559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.279579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.279893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.279911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.280240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.280258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.280592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.280609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.280907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.280925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.281166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.281182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.281512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.281530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.281878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.281896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.282222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.783 [2024-11-26 19:26:58.282239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.783 qpair failed and we were unable to recover it. 00:30:45.783 [2024-11-26 19:26:58.282563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.282580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.282900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.282919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.283272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.283290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.283613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.283631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.283955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.283973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.284292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.284311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.284643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.284661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.284801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.284818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.285155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.285173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.285486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.285504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.285839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.285858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.286150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.286167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.286417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.286434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.286743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.286762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.287084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.287102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.287413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.287431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.287770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.287787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.287999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.288016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.288417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.288435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.288771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.288788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.288964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.288982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.289330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.289348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.289681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.289698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.290035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.290054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.290378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.290397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.290709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.290727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.291043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.291062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.291365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.291382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.291600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.291618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.291950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.291968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.292300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.292317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.292642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.292664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.292994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.293014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.293326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.293343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.293679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.784 [2024-11-26 19:26:58.293696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.784 qpair failed and we were unable to recover it. 00:30:45.784 [2024-11-26 19:26:58.294008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.294025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.294203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.294223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.294542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.294559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.294892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.294909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.295279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.295296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.295615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.295631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.295947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.295965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.296296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.296315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.296644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.296661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.296993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.297011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.297334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.297352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.297687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.297706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.298029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.298047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.298383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.298401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.298717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.298734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.299063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.299081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.299392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.299409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.299734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.299752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.299936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.299955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.300291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.300309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.300642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.300660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.300986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.301005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.301355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.301374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.301706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.301723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.302065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.302085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.302400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.302418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.302768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.302787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.303163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.303181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.303486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.303504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.303828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.303846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.304187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.304206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.304539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.304557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.304889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.304907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.305095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.785 [2024-11-26 19:26:58.305114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.785 qpair failed and we were unable to recover it. 00:30:45.785 [2024-11-26 19:26:58.305449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.305467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.305684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.305702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.306012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.306037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.306373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.306391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.306691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.306708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.307041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.307060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.307365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.307382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.307697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.307715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.308033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.308051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.308410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.308428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.308763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.308781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.309001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.309020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.309371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.309390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.309698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.309716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.310053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.310073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.310387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.310404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.310726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.310745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.311052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.311070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.311389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.311406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.311718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.311736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.312044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.312061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.312401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.312418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.312764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.312783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.313178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.313196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.313512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.313530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.313734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.313752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.314091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.314109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.314414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.314431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.314747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.314764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.315117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.315135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.315460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.786 [2024-11-26 19:26:58.315479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.786 qpair failed and we were unable to recover it. 00:30:45.786 [2024-11-26 19:26:58.315809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.315826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.316145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.316163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.316336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.316354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.316686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.316703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.317077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.317096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.317413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.317432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.317764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.317782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.318118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.318136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.318459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.318476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.318819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.318837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.319144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.319162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.319491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.319510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.319820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.319837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.320150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.320168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.320504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.320521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.320856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.320884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.321197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.321215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.321532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.321550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.321889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.321907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.322275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.322294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.322627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.322644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.322814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.322831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.323123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.323144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.323469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.323486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.323834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.323853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.324167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.324184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.324526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.324544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.787 qpair failed and we were unable to recover it. 00:30:45.787 [2024-11-26 19:26:58.324876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.787 [2024-11-26 19:26:58.324894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.325241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.325257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.325598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.325616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.325955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.325973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.326311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.326329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.326664] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.326683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.326906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.326924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.327269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.327288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.327614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.327631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.327980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.327997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.328320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.328337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.328401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.328420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.328712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.328728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.329036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.329055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.329401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.329420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.329734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.329752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.330097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.330115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.330473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.330492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.330819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.330836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.331219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.331238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.331529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.331547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.331879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.331897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.332118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.332135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.332476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.332495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.332833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.332851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.333194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.333213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.333534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.333551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.333773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.333790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.334016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.334034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.334368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.334385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.334581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.334598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.334920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.334938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.335281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.788 [2024-11-26 19:26:58.335300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.788 qpair failed and we were unable to recover it. 00:30:45.788 [2024-11-26 19:26:58.335626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.335644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.335974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.335991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.336376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.336393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.336700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.336717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.337045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.337063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.337400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.337420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.337770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.337788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.338104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.338121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.338426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.338444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.338623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.338640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.338968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.338987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.339331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.339348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.339752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.339769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.340033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.340050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.340376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.340394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.340734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.340751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.341088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.341107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.341427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.341444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.341636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.341658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.341997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.342017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.342393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.342410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.342721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.342737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.343087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.343106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.343443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.343462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.343776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.343793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.344096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.344115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.344430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.344447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.344782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.344800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.345126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.345144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.345480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.345498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.345814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.345832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:45.789 [2024-11-26 19:26:58.346131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:45.789 [2024-11-26 19:26:58.346149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:45.789 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.346490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.346513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.101 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.346845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.346879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.101 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.347240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.347258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.101 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.347592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.347610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.101 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.347937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.347955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.101 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.348303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.348319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.101 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.348638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.348656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.101 qpair failed and we were unable to recover it. 00:30:46.101 [2024-11-26 19:26:58.348995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.101 [2024-11-26 19:26:58.349013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.349353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.349370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.349695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.349711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.350031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.350050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.350381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.350400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.350593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.350612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.350940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.350957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.351280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.351298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.351633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.351649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.351987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.352005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.352339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.352356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.352679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.352697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.353031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.353051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.353385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.353403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.353727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.353744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.354051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.354069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.354453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.354472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.354775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.354791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.355138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.355156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.355382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.355403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.355782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.355800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.356137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.356156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.356472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.356491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.356721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.356739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.357049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.357067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.357375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.357391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.357718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.357736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.358086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.358103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.358478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.358496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.358829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.358847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.359171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.359188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.359566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.359585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.359917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.359936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.360292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.360311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.360532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.360549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.360771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.360790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.361145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.361163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.361496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.361512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.102 qpair failed and we were unable to recover it. 00:30:46.102 [2024-11-26 19:26:58.361826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.102 [2024-11-26 19:26:58.361844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.362165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.362184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.362503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.362522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.362888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.362907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.363244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.363261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.363583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.363599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.363948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.363966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.364295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.364314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.364651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.364670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.365009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.365026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.365443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.365461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.365792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.365809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.366047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.366064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.366384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.366402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.366744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.366762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.367103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.367123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.367451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.367469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.367846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.367874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.368125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.368141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.368470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.368487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.368844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.368873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.369493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.369524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.369847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.369879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.370267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.370286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.370625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.370642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.370880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.370899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.371350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.371368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.371703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.371720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.372171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.372237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.372620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.372642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.372937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.372957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.373281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.373299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.373615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.373633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.373960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.373978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.374366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.374384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.103 [2024-11-26 19:26:58.374671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.103 [2024-11-26 19:26:58.374688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.103 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.375028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.375047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.375401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.375421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.375751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.375768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.376095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.376113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.376311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.376333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.376679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.376697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.377047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.377066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.377404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.377422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.377749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.377767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.377977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.377996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.378357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.378376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.378706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.378723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.379052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.379073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.379369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.379389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.379781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.379801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.380110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.380132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.380459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.380477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.380854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.380906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.381258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.381276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.381615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.381635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.381836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.381856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.382269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.382287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.382626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.382644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.382983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.383002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.383358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.383379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.383705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.383726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.384043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.384060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.384444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.384463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.384794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.384813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.385152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.385171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.385515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.385533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.385853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.385881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.386263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.386283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.386619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.386636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.386896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.386915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.387256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.387273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.387606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.387624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.104 [2024-11-26 19:26:58.387944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.104 [2024-11-26 19:26:58.387961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.104 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.388340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.388358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.388699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.388716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.389057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.389076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.389432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.389452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.389769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.389789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.390126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.390146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.390470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.390489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.390806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.390823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.391139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.391160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.391486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.391502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.391936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.391956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.392297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.392316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.392613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.392631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.392844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.392860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.393172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.393191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.393556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.393573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.393908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.393928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.394278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.394297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.394628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.394646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.394859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.394886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.395217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.395236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.395603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.395621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.395952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.395970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.396182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.396202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.396544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.396565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.396893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.396913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.397225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.397242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.397565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.397586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.397831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.397847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.398212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.398230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.398546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.398563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.398881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.398902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.399246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.399264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.399581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.399598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.399931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.399950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.400302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.400322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.400656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.400676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.105 qpair failed and we were unable to recover it. 00:30:46.105 [2024-11-26 19:26:58.400895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.105 [2024-11-26 19:26:58.400914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.401263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.401281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.401688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.401707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.402042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.402061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.402402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.402420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.402751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.402769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.403120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.403140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.403379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.403395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.403769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.403787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.404004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.404023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.404342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.404361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.404670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.404687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.405023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.405042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.405387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.405406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.405730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.405748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.406088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.406107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.406448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.406465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.406799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.406818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.407143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.407161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.407495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.407515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.407924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.407945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.408159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.408177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.408505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.408525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.408872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.408892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.409237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.409256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.409460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.409479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.409815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.409832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.410036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.410057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.410388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.410407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.410717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.410735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.411075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.411098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.411430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.106 [2024-11-26 19:26:58.411448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.106 qpair failed and we were unable to recover it. 00:30:46.106 [2024-11-26 19:26:58.411784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.411803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.412147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.412169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.412385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.412407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.412752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.412773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.413113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.413133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.413436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.413455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.413775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.413793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.414209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.414229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.414562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.414580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.414907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.414925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.415252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.415272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.415503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.415521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.415909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.415928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.416260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.416278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.416486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.416504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.416838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.416855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.417208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.417225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.417556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.417575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.417896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.417915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.418279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.418297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.418606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.418622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.418987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.419007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.419327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.419344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.419651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.419670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.419891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.419912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.420217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.420238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.420558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.420577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.420764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.420783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.421133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.421152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.421484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.421500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.421826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.421844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.422158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.422176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.422390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.422410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.422745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.422764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.422962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.422984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.423336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.423356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.423676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.423695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.424025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.107 [2024-11-26 19:26:58.424046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.107 qpair failed and we were unable to recover it. 00:30:46.107 [2024-11-26 19:26:58.424373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.424394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.424735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.424753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.425045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.425063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.425403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.425421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.425742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.425760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.426098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.426117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.426308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.426325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.426659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.426678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.427024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.427043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.427377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.427397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.427732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.427751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.428087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.428106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.428461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.428479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.428818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.428836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.429187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.429207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.429536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.429556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.429887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.429907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.430130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.430150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.430472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.430489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.430797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.430816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.431060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.431078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.431399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.431418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.431752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.431771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.432013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.432033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.432395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.432412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.432629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.432648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.433003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.433023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.433386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.433405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.433729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.433748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.434114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.434133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.434466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.434485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.434815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.434835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.435151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.435168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.435504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.435523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.435835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.435853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.436185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.436205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.436524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.108 [2024-11-26 19:26:58.436542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.108 qpair failed and we were unable to recover it. 00:30:46.108 [2024-11-26 19:26:58.436880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.436901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.437241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.437261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.437580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.437599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.437910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.437935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.438350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.438369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.438699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.438719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.439056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.439075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.439431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.439449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.439786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.439803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.440143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.440162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.440480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.440496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.440821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.440842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.441190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.441209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.441548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.441569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.441914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.441933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.442146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.442164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.442512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.442530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.442717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.442738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.443086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.443106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.443427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.443445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.443783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.443802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.444114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.444134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.444471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.444489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.444707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.444726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.445036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.445056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.445393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.445410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.445739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.445758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.446154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.446174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.446482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.446500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.446869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.446889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.447090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.447108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.447436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.447456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.447797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.447815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.448157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.448177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.448508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.448526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.448851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.109 [2024-11-26 19:26:58.448880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.109 qpair failed and we were unable to recover it. 00:30:46.109 [2024-11-26 19:26:58.449095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.449112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.449516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.449536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.449885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.449905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.450235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.450255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.450593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.450614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.450952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.450972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.451318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.451339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.451653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.451677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.452009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.452029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.452389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.452406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.452742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.452760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.453084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.453104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.453446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.453466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.453796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.453814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.454151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.454171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.454498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.454517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.454888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.454909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.455133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.455152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.455464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.455482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.455710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.455729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.456085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.456104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.456484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.456501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.456820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.456837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.457178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.457197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.457553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.457572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.457909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.457928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.458277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.458296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.458622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.458640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.458975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.458993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.459202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.459220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.459454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.459472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.459808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.459827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.460153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.460171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.460530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.460549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.460880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.460900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.461251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.461269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.461607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.461625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.461846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.110 [2024-11-26 19:26:58.461873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.110 qpair failed and we were unable to recover it. 00:30:46.110 [2024-11-26 19:26:58.462122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.462139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.462491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.462509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.462845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.462876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.463177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.463195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.463533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.463552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.463878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.463897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.464249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.464268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.464597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.464615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.464805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.464822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.465165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.465188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.465520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.465539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.465853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.465885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.466220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.466239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.466565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.466582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.466919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.466936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.467276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.467293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.467607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.467624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.467811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.467831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.468170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.468189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.468550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.468569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.468770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.468789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.469028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.469046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.469402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.469420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.469747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.469766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.470100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.470120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.470440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.470459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.470806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.470825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.471147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.471167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.471495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.471514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.471836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.471856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.472214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.472233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.472567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.472586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.111 qpair failed and we were unable to recover it. 00:30:46.111 [2024-11-26 19:26:58.472916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.111 [2024-11-26 19:26:58.472936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.473291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.473309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.473650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.473668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.473988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.474006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.474323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.474340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.474658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.474676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.474878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.474896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.475270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.475287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.475604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.475623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.475947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.475966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.476309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.476328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.476675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.476693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.477066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.477085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.477417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.477435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.477787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.477805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.478125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.478143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.478479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.478498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.478818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.478840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.479204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.479223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.479545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.479562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.479888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.479908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.480258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.480276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.480592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.480609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.480949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.480967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.481313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.481333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.481653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.481671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.481993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.482011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.482327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.482345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.482690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.482709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.483034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.483052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.483409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.483427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.483759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.483776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.484123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.484143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.484496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.484514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.484830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.484847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.485205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.485223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.485456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.485472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.485790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.485808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.112 [2024-11-26 19:26:58.486155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.112 [2024-11-26 19:26:58.486174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.112 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.486510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.486528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.486873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.486892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.487226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.487244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.487559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.487576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.487907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.487925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.488259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.488279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.488599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.488618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.488930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.488948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.489298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.489316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.489649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.489666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.490021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.490040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.490401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.490419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.490742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.490759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.491088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.491106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.491434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.491454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.491767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.491785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.492129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.492146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.492471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.492489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.492844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.492868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.493111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.493129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.493312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.493329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.493630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.493648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.494009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.494026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.494371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.494389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.494735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.494754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.495101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.495119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.495472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.495491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.495830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.495849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.496183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.496201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.496551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.496570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.496882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.496902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.497112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.497130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.497471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.497489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.497822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.497840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.498217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.498236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.498570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.498589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.498905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.498923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.113 [2024-11-26 19:26:58.499280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.113 [2024-11-26 19:26:58.499299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.113 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.499621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.499638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.499983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.500000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.500369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.500386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.500731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.500750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.501125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.501144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.501476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.501495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.501828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.501845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.502158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.502179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.502583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.502601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.502941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.502961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.503303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.503321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.503654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.503673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.503999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.504018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.504335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.504355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.504679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.504697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.504921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.504938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.505255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.505273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.505610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.505630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.505832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.505849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.506209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.506228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.506554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.506572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.506910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.506929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.507267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.507286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.507665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.507682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.508041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.508058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.508261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.508278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.508484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.508502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.508837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.508855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.509184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.509202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.509554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.509571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.509919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.509938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.510265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.510283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.510593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.510609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.510889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.510907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.511260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.511279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.511606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.511626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.511962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.511981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.512329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.114 [2024-11-26 19:26:58.512347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.114 qpair failed and we were unable to recover it. 00:30:46.114 [2024-11-26 19:26:58.512672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.512690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.513029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.513047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.513392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.513410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.513751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.513770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.514106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.514128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.514457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.514477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.514797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.514815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.515143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.515163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.515387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.515404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.515739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.515765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.516088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.516106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.516448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.516467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.516683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.516702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.517032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.517049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.517406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.517424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.517766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.517785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.518128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.518147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.518495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.518513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.518836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.518854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.519168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.519187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.519491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.519510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.519884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.519904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.520223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.520241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.520571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.520589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.520924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.520942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.521285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.521304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.521641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.521659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.521871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.521889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.522214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.522232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.522453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.522469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.522812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.522830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.523214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.523232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.523573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.523590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.115 [2024-11-26 19:26:58.523960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.115 [2024-11-26 19:26:58.523978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.115 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.524209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.524226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.524607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.524624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.524953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.524973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.525366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.525384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.525704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.525723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.526103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.526122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.526487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.526505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.526858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.526883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.527235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.527252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.527589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.527607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.527801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.527822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.528142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.528160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.528501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.528519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.528856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.528883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.529178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.529195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.529394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.529415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.529732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.529750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.530090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.530110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.530442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.530459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.530793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.530811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.531131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.531150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.531480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.531499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.531717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.531735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.531969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.531987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.532328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.532346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.532728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.532745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.533093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.533112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.533429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.533447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.533806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.533824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.534134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.534152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.534494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.534512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.534742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.534760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.535033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.535052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.535443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.535462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.535784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.535802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.536142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.536160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.536481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.536501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.536836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.116 [2024-11-26 19:26:58.536855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.116 qpair failed and we were unable to recover it. 00:30:46.116 [2024-11-26 19:26:58.537188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.537206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.537544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.537563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.537889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.537909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.538234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.538252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.538578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.538595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.538938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.538957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.539292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.539311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.539667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.539684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.540010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.540029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.540375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.540392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.540712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.540731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.540942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.540961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.541316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.541335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.541665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.541683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.542022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.542042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.542373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.542392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.542732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.542751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.543131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.543153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.543471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.543489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.543731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.543748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.544080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.544097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.544442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.544461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.544681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.544697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.545029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.545048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.545395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.545412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.545761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.545780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.546120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.546139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.546478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.546498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.546841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.546861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.547076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.547094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.547415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.547432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.547752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.547771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.548109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.548129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.548320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.548340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.548676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.548695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.549034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.549053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.549382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.549401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.549733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.117 [2024-11-26 19:26:58.549751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.117 qpair failed and we were unable to recover it. 00:30:46.117 [2024-11-26 19:26:58.550087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.550105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.550476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.550493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.550824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.550840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.551055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.551075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.551410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.551429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.551767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.551785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.552142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.552162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.552525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.552542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.552889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.552908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.553105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.553123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.553458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.553478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.553812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.553829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.554154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.554174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.554510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.554528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.554871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.554890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.555202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.555219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.555545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.555564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.555883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.555901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.556240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.556258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.556587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.556608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.557005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.557024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.557363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.557380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.557560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.557578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.557922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.557941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.558299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.558318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.558654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.558673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.559013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.559032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.559374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.559393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.559793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.559810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.560109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.560127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.560466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.560484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.560872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.560891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.561201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.561218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.561435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.561454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.561784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.561802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.562148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.562167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.562535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.562553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.118 [2024-11-26 19:26:58.562793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.118 [2024-11-26 19:26:58.562810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.118 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.563186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.563204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.563537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.563556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.563881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.563900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.564236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.564254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.564602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.564620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.564984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.565004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.565334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.565352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.565689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.565708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.566040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.566058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.566411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.566430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.566653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.566673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.567143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.567162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.567499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.567517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.567846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.567869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.568196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.568215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.568552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.568571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.568756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.568773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.569077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.569097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.569445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.569463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.569781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.569800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.570138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.570157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.570552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.570574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.570906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.570924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.571322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.571340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.571669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.571688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.572023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.572041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.572394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.572413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.572629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.572650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.572976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.572994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.573334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.573352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.573538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.573556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.573799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.573817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.574126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.574145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.574473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.574490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.574809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.574828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.575186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.575204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.575552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.575571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.119 [2024-11-26 19:26:58.575903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.119 [2024-11-26 19:26:58.575922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.119 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.576292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.576310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.576655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.576675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.577017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.577035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.577352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.577371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.577727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.577746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.578083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.578101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.578322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.578342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.578681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.578700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.579054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.579073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.579429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.579448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.579667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.579686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.580030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.580047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.580379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.580397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.580737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.580757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.581093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.581112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.581419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.581436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.581755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.581772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.582116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.582136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.582458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.582476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.582787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.582805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.583141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.583160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.583473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.583491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.583807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.583827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.584043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.584068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.584429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.584448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.584793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.584811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.585029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.585048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.585399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.585420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.585744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.585764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.586090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.586110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.586443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.586462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.586791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.586811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.587173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.120 [2024-11-26 19:26:58.587193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.120 qpair failed and we were unable to recover it. 00:30:46.120 [2024-11-26 19:26:58.587525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.587544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.587701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.587722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.588030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.588049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.588435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.588452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.588801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.588820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.589152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.589170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.589508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.589527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.589879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.589898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.590236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.590255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.590569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.590587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.590912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.590929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.591330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.591347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.591694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.591712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.592031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.592050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.592254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.592272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.592470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.592488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.592843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.592860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.593241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.593259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.593595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.593614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.593946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.593963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.594298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.594316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.594628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.594646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.594992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.595010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.595369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.595388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.595605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.595626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.595958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.595976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.596311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.596330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.596648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.596666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.597083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.597102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.597427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.597446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.597786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.597808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.598143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.598162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.598496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.598514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.598836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.598855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.599178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.599196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.599544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.599562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.599931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.599949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.600286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.121 [2024-11-26 19:26:58.600305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.121 qpair failed and we were unable to recover it. 00:30:46.121 [2024-11-26 19:26:58.600635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.600653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.600978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.600996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.601358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.601375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.601701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.601720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.602068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.602086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.602411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.602427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.602775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.602792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.602915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.602932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.603311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.603329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.603659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.603676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.603886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.603905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.604220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.604237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.604574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.604592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.604910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.604929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.605269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.605286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.605628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.605646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.605963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.605981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.606351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.606369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.606678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.606695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.607035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.607053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.607398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.607417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.607740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.607757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.607975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.607993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.608332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.608350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.608672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.608690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.609065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.609084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.609423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.609442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.609763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.609780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.610117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.610134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.610456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.610473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.610811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.610830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.611209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.611227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.611563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.611586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.611925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.611944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.612274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.612293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.612608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.612626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.612943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.612962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.613316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.122 [2024-11-26 19:26:58.613335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.122 qpair failed and we were unable to recover it. 00:30:46.122 [2024-11-26 19:26:58.613526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.613546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.613886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.613905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.614243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.614262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.614583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.614601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.614940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.614957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.615298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.615316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.615600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.615617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.615945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.615964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.616184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.616200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.616541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.616559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.616876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.616893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.617282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.617299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.617638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.617657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.617990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.618008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.618360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.618379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.618744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.618761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.619003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.619020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.619417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.619436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.619774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.619792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.620129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.620148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.620457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.620476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.620805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.620824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.621219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.621238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.621589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.621607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.621943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.621961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.622308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.622327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.622650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.622668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.622854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.622879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.623236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.623255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.623570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.623589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.623931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.623950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.624276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.624295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.624514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.624532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.624846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.624869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.123 qpair failed and we were unable to recover it. 00:30:46.123 [2024-11-26 19:26:58.625063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.123 [2024-11-26 19:26:58.625086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.625415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.625435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.625776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.625794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.626137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.626154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.626484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.626503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.626825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.626845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.627183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.627201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.627527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.627546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.627883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.627902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.628246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.628265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.628613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.628631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.628965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.628983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.629317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.629337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.629660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.629680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.629872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.629893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.630230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.630251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.630634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.630653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.630985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.631004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.631210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.631232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.631542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.631561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.631940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.631961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.632281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.632300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.632627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.632645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.632980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.632999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.633329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.633349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.633682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.633702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.634039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.634059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.634391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.634413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.634742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.634762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.635112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.635132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.635466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.635486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.635822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.635842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.636161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.636180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.636380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.124 [2024-11-26 19:26:58.636401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.124 qpair failed and we were unable to recover it. 00:30:46.124 [2024-11-26 19:26:58.636753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.636773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.637089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.637108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.637438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.637457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.637786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.637806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.638041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.638063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.638419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.638440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.638768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.638791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.639149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.639170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.639509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.639530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.639880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.639902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.640230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.640248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.640586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.640607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.640812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.640830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.641205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.641224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.641563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.641581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.641966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.641999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.642350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.642369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.642576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.642596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.642951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.642970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.643346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.643364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.643706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.643725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.644084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.644102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.644437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.644456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.644777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.644797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.645179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.645198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.645524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.645544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.645877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.645898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.646268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.646286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.646637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.646656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.646841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.646872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.647231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.647250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.647428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.647446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.647806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.647826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.648142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.648161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.648487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.648507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.648830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.648847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.649168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.649187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.649518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.125 [2024-11-26 19:26:58.649537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.125 qpair failed and we were unable to recover it. 00:30:46.125 [2024-11-26 19:26:58.649889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.649909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.650255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.650274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.650601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.650620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.650982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.651001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.651310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.651329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.651677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.651698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.652034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.652055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.652386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.652405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.652758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.652784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.653136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.653156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.653491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.653510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.653844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.653872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.654231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.654251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.654646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.654663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.654972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.654990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.655341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.655360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.655741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.655760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.656099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.656118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.656473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.656491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.656710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.656728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.656943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.656963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.657308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.657327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.657662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.657684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.657900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.657920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.658267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.658286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.658613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.658633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.658974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.658992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.659305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.659325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.659683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.659705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.660033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.660054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.660250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.660270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.660610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.660628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.660970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.660990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.661302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.661321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.661668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.661687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.662025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.662045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.662349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.662365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.662689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.126 [2024-11-26 19:26:58.662710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.126 qpair failed and we were unable to recover it. 00:30:46.126 [2024-11-26 19:26:58.662938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.662958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.663269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.663286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.663619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.663638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.663952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.663970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.664316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.664335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.664655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.664674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.665012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.665031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.665353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.665371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.665716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.665734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.666094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.666114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.666442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.666460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.666788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.666808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.667116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.667135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.667457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.667476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.667802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.667822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.668156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.668177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.668511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.668531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.668860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.668887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.669222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.669241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.669596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.669613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.669955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.669974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.670317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.670336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.670650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.670669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.670980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.670999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.671343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.671362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.671700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.671719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.672056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.672075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.672396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.672414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.672782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.672801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.673113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.673132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.673467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.673487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.673806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.673825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.674136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.674156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.674472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.674491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.674822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.674843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.675231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.675251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.675606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.675626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.675946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.675969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.127 qpair failed and we were unable to recover it. 00:30:46.127 [2024-11-26 19:26:58.676153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.127 [2024-11-26 19:26:58.676171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.676501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.676519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.676860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.676889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.677112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.677133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.677467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.677487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.677814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.677833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.678180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.678199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.678607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.678626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.678932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.678950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.679330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.679348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.679689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.679708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.680056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.680077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.680410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.680430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.680752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.680772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.681124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.681144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.681469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.681487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.681880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.681900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.682246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.682266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.682614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.682633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.682940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.682959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.683311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.683328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.683652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.683671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.684009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.684028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.684237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.684256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.684609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.684627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.684952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.684971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.685327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.685346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.685724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.685742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.686078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.686098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.686422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.686442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.686781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.686797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.687149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.687168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.687512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.687531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.687859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.687890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.688206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.688226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.688547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.688566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.688747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.688767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.689111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.128 [2024-11-26 19:26:58.689130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.128 qpair failed and we were unable to recover it. 00:30:46.128 [2024-11-26 19:26:58.689496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.689514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.689847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.689895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.690262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.690280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.690615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.690633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.690944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.690963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.691210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.691226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.691565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.691584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.691801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.691823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.692161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.692183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.692514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.692533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.692877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.692899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.693240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.693259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.693597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.693615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.693991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.694011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.694354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.694373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.694689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.694707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.695041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.695061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.695377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.695394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.695734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.695753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.696075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.696094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.696431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.696450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.696787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.696804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.697027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.697044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.697364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.697382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.697739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.697758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.698104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.698122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.698352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.698369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.698729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.698747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.699128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.699146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.699330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.699347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.699643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.699662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.699981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.699999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.700311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.700329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.129 [2024-11-26 19:26:58.700656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.129 [2024-11-26 19:26:58.700675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.129 qpair failed and we were unable to recover it. 00:30:46.130 [2024-11-26 19:26:58.700969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.130 [2024-11-26 19:26:58.700987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.130 qpair failed and we were unable to recover it. 00:30:46.130 [2024-11-26 19:26:58.701330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.130 [2024-11-26 19:26:58.701348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.130 qpair failed and we were unable to recover it. 00:30:46.130 [2024-11-26 19:26:58.701700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.130 [2024-11-26 19:26:58.701718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.130 qpair failed and we were unable to recover it. 00:30:46.407 [2024-11-26 19:26:58.702062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.407 [2024-11-26 19:26:58.702086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.407 qpair failed and we were unable to recover it. 00:30:46.407 [2024-11-26 19:26:58.702453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.407 [2024-11-26 19:26:58.702473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.407 qpair failed and we were unable to recover it. 00:30:46.407 [2024-11-26 19:26:58.702797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.407 [2024-11-26 19:26:58.702814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.407 qpair failed and we were unable to recover it. 00:30:46.407 [2024-11-26 19:26:58.703157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.407 [2024-11-26 19:26:58.703175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.407 qpair failed and we were unable to recover it. 00:30:46.407 [2024-11-26 19:26:58.703497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.407 [2024-11-26 19:26:58.703517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.407 qpair failed and we were unable to recover it. 00:30:46.407 [2024-11-26 19:26:58.703828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.703846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.704214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.704233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.704570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.704587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.704921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.704939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.705287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.705305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.705714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.705733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.706039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.706056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.706380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.706397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.706753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.706772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.707114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.707132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.707451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.707467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.707815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.707832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.708248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.708267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.708581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.708599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.708942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.708960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.709292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.709312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.709651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.709669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.709855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.709884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.710224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.710242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.710578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.710597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.710935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.710953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.711298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.711317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.711659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.711676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.712012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.712029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.712263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.712281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.712481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.712499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.712786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.712804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.713120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.713137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.713369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.713389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.713699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.713716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.714089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.714107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.714469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.714487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.714826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.714842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.715163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.715181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.715515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.715532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.715885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.715904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.716244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.716261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.408 [2024-11-26 19:26:58.716671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.408 [2024-11-26 19:26:58.716688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.408 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.717030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.717050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.717248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.717270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.717600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.717620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.717991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.718009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.719780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.719826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.720182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.720205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.720503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.720521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.720843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.720873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.721219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.721238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.721572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.721591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.721932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.721954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.722284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.722304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.722630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.722648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.722889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.722907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.723222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.723242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.723571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.723589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.723929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.723949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.724288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.724309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.724641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.724658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.726249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.726293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.726669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.726690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.727023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.727041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.727388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.727408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.727734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.727752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.728097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.728115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.728318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.728338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.728661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.728679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.728994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.729011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.729351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.729369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.729666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.729682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.730040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.730058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.730386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.730404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.730751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.730769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.731108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.731125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.731490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.731507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.731836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.731854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.732199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.732217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.409 [2024-11-26 19:26:58.732545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.409 [2024-11-26 19:26:58.732562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.409 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.732901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.732919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.733294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.733313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.733691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.733709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.734039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.734061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.734440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.734458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.734852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.734877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.735236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.735255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.735590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.735608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.735948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.735966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.736306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.736323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.736660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.736679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.737020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.737038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.737374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.737393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.737718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.737736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.737931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.737951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.738291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.738308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.738639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.738658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.738986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.739006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.739343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.739362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.739699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.739717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.740057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.740077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.740402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.740419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.740637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.740653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.740988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.741006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.741348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.741368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.741689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.741707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.741923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.741940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.742276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.742295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.742625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.742641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.742998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.743016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.743365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.743384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.743748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.743765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.743981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.744000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.744360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.744378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.744716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.744735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.745043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.745061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.745395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.745414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.410 [2024-11-26 19:26:58.745803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.410 [2024-11-26 19:26:58.745821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.410 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.746171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.746190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.746520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.746539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.746878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.746897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.747273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.747291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.747642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.747660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.747944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.747966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.748181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.748198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.748459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.748478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.748816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.748834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.749170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.749190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.749498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.749515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.749831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.749849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.750184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.750203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.750417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.750435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.750753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.750770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.751121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.751138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.751471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.751489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.751806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.751823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.752170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.752189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.752529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.752548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.752885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.752904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.754874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.754911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.755307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.755329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.755644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.755662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.756004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.756022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.757041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.757076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.757425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.757446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.757763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.757781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.758131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.758149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.758488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.758507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.758840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.758859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.759196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.759215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.759540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.759559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.759897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.759918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.760297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.760315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.761437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.761475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.761822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.761843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.411 [2024-11-26 19:26:58.762225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.411 [2024-11-26 19:26:58.762244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.411 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.762598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.762618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.763675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.763712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.763938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.763959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.764201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.764220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.764574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.764592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.764921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.764939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.765286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.765305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.765644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.765667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.765889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.765912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.766273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.766293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.766506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.766524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.766872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.766891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.767210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.767230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.768252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.768293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.768685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.768706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.769030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.769050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.769393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.769412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.769741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.769758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.770103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.770123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.770497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.770516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.770844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.770873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.771204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.771222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.772286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.772325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.772672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.772693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.773018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.773037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.773368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.773388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.773622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.773640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.774564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.774598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.774820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.774838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.775171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.775190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.775526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.775545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.775889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.412 [2024-11-26 19:26:58.775909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.412 qpair failed and we were unable to recover it. 00:30:46.412 [2024-11-26 19:26:58.776212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.776230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.776544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.776563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.776902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.776923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.777255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.777272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.777506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.777524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.777738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.777757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.777990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.778009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.778346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.778363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.778745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.778762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.778988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.779004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.779312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.779330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.779651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.779670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.779989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.780008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.780272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.780289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.780632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.780651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.780987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.781010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.781360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.781378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.781542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.781560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.781930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.781949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.782292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.782308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.782644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.782661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.782902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.782919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.783262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.783281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.783619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.783639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.783960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.783978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.784313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.784330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.784555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.784573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.784919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.784939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.785166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.785185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.785530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.785550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.785903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.785922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.786125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.786143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.786462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.786480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.786804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.786821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.787190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.787209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.787537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.787556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.787888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.787906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.788111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.413 [2024-11-26 19:26:58.788128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.413 qpair failed and we were unable to recover it. 00:30:46.413 [2024-11-26 19:26:58.788467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.788486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.788818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.788837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.789188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.789208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.789555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.789574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.789913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.789933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.790294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.790313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.790657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.790675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.790914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.790932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.791266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.791283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.791612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.791629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.791944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.791962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.792322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.792340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.792698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.792716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.793056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.793074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.793290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.793309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.793616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.793633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.793984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.794004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.794251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.794272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.794581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.794599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.794932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.794951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.795267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.795285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.795636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.795654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.795996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.796015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.796351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.796368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.796711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.796729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.797060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.797078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.797472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.797489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.797808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.797825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.798031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.798050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.798380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.798399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.798760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.798779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.799204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.799222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.799562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.799582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.799900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.799920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.800272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.800290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.800624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.800643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.800989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.801007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.414 [2024-11-26 19:26:58.801346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.414 [2024-11-26 19:26:58.801366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.414 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.801546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.801566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.801969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.801987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.802382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.802400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.802733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.802752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.803093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.803112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.803438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.803458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.803812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.803830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.804016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.804033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.804337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.804355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.804686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.804706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.805026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.805044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.805393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.805413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.805751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.805769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.805998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.806015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.806344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.806363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.806694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.806713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.806923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.806944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.807318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.807338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.807602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.807620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.807941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.807962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.808177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.808194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.808568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.808587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.808933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.808951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.809307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.809325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.809511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.809530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.809745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.809761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.809993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.810012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.810386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.810403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.810737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.810756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.811049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.811067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.811437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.811455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.811648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.811666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.811991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.812009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.812350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.812368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.812716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.812733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.813075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.813096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.813316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.813334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.415 [2024-11-26 19:26:58.813672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.415 [2024-11-26 19:26:58.813693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.415 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.813931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.813950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.814297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.814316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.814648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.814666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.815016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.815035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.815241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.815259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.815596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.815615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.815950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.815968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.816327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.816346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.816583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.816603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.816803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.816821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.817166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.817185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.817506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.817526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.817861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.817893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.818238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.818256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.818613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.818630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.818992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.819013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.819321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.819339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.819683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.819702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.820029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.820046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.820298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.820315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.820649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.820667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.820886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.820909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.821167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.821183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.821529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.821547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.821888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.821908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.822231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.822249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.822606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.822623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.822956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.822974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.823207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.823225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.823437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.823454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.823770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.823788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.824159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.824176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.824537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.824555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.824875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.824894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.416 qpair failed and we were unable to recover it. 00:30:46.416 [2024-11-26 19:26:58.825213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.416 [2024-11-26 19:26:58.825231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.825631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.825647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.825954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.825971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.826320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.826338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.826534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.826552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.826743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.826760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.827118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.827137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.827474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.827490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.827716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.827733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.827994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.828011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.828296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.828313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.828718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.828735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.828934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.828954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.829297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.829315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.829535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.829551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.829881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.829900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.830230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.830248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.830574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.830593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.830937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.830954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.831290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.831309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.831626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.831643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.831939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.831956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.832316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.832333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.832709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.832726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.833049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.833066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.833401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.833418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.833745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.833763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.834093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.834116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.834337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.834355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.834563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.834584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.834922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.834940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.835330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.835347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.835657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.835676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.835994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.836012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.836330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.836348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.836688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.836705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.837023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.837043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.417 qpair failed and we were unable to recover it. 00:30:46.417 [2024-11-26 19:26:58.837368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.417 [2024-11-26 19:26:58.837385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.837616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.837635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.837835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.837855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.838236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.838253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.838453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.838471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.838878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.838898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.839265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.839282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.839617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.839635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.839988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.840006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.840338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.840356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.840690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.840709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.841048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.841065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.841385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.841403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.841760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.841778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.842097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.842118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.842446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.842464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.842785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.842804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.843217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.843235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.843582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.843602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.843941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.843958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.844210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.844226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.844543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.844562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.844789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.844807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.845225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.845243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.845598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.845617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.845950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.845968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.846328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.846347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.846691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.846709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.847048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.847068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.847403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.847420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.847636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.847655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.847892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.847912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.848260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.848277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.848622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.848641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.848880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.848898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.849112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.849130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.849489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.849507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.849738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.849755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.418 qpair failed and we were unable to recover it. 00:30:46.418 [2024-11-26 19:26:58.850133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.418 [2024-11-26 19:26:58.850152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.850563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.850582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.850777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.850794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.851103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.851122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.851453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.851470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.851779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.851797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.852156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.852175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.852536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.852554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.852774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.852791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.853129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.853147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.853480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.853498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.853723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.853743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.854117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.854136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.854477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.854495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.854729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.854747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.855082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.855101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.855435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.855454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.855763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.855781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.856072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.856091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.856433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.856455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.856773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.856793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.857185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.857204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.857516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.857535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.857745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.857764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.857895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.857915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.858264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.858283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.858623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.858644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.858992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.859010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.859385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.859403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.859745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.859764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.860156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.860175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.860517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.860534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.860869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.860887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.861288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.861305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.861534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.861551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.861872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.861890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.862229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.862248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.862576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.862594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.419 [2024-11-26 19:26:58.862994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.419 [2024-11-26 19:26:58.863012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.419 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.863375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.863394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.863728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.863745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.864064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.864083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.864420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.864438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.864813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.864831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.865060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.865077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.865411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.865428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.865753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.865773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.866092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.866111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.866452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.866471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.866778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.866797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.867145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.867164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.867502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.867520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.867853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.867888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.868231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.868249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.868574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.868590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.868772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.868791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.868992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.869010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.869358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.869376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.869719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.869738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.870058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.870080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.870403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.870422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.870764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.870782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.871119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.871139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.871483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.871500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.871687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.871704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.871962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.871981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.872314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.872331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.872715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.872733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.873160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.873179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.873484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.873502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.873843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.873868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.874179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.874197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.874528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.874548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.874735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.874753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.875115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.875133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.875341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.875360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.420 [2024-11-26 19:26:58.875606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.420 [2024-11-26 19:26:58.875625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.420 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.875992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.876011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.876426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.876445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.876668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.876688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.876929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.876949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.877322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.877340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.877659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.877679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.878006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.878025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.878378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.878397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.878757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.878775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.879167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.879188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.879551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.879568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.879914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.879932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.880273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.880291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.880429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.880446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.880640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.880658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.881010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.881029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.881377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.881396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.881752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.881771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.882004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.882021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.882267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.882285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.882638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.882656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.882993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.883014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.883338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.883359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.883696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.883715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.884046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.884067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.884279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.884298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.884611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.884629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.885031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.885051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.885348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.885368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.885691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.885712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.886044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.886063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.886248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.886266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.886623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.886642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.886992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.421 [2024-11-26 19:26:58.887010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.421 qpair failed and we were unable to recover it. 00:30:46.421 [2024-11-26 19:26:58.887358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.887376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.887755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.887775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.888204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.888223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.888547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.888565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.888907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.888925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.889285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.889302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.889641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.889660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.889986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.890004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.890365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.890384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.890621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.890640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.890943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.890961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.891335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.891352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.891462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.891479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.891766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.891786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.892116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.892134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.892354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.892372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.892697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.892716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.893082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.893103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.893438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.893456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.893780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.893798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.894083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.894101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.894353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.894371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.894692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.894710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.895137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.895156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.895424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.895442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.895791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.895809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.896152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.896172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.896513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.896532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.896707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.896731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.896992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.897013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.897332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.897351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.897538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.897558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.897885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.897904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.898270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.898287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.898638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.898657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.898925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.898946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.899292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.899312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.899643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.422 [2024-11-26 19:26:58.899660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.422 qpair failed and we were unable to recover it. 00:30:46.422 [2024-11-26 19:26:58.900003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.900021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.900395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.900413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.900743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.900760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.901122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.901142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.901463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.901484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.901805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.901825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.902166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.902185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.902568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.902586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.902800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.902820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.903166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.903186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.903507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.903525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.903748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.903772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.904096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.904115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.904338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.904358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.904671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.904688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.905034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.905054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.905353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.905372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.905708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.905726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.906066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.906085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.906430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.906448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.906774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.906792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.906999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.907019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.907396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.907415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.907714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.907732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.907852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.907889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.908190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.908209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.908553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.908574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.908918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.908939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.909302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.909321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.909651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.909668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.909896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.909918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.910217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.910237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.910564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.910582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.910910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.910928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.911264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.911283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.911595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.911615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.911936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.911956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.912177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.423 [2024-11-26 19:26:58.912197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.423 qpair failed and we were unable to recover it. 00:30:46.423 [2024-11-26 19:26:58.912523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.912542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.912782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.912800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.913106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.913123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.913441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.913462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.913797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.913817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.914128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.914148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.914499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.914518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.914792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.914810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.915164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.915183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.915519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.915537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.915876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.915894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.916258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.916279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.916490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.916507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.916890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.916909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.917263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.917283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.917627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.917648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.917979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.917998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.918378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.918396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.918739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.918758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.918995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.919012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.919421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.919441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.919790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.919810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.920132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.920151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.920485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.920505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.920818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.920838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.921175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.921195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.921538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.921555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.921917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.921935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.922305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.922324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.922711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.922729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.923049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.923069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.923330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.923347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.923679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.923701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.924046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.924066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.924420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.924438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.924769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.924789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.925143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.925163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.925363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.424 [2024-11-26 19:26:58.925383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.424 qpair failed and we were unable to recover it. 00:30:46.424 [2024-11-26 19:26:58.925639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.925661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.926012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.926032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.926373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.926392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.926779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.926798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.927127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.927146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.927467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.927487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.927838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.927856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.928186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.928207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.928412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.928433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.928656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.928675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.929008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.929028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.929377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.929397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.929742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.929761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.930114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.930137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.930455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.930473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.930795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.930813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.931169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.931190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.931384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.931403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.931643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.931661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.932010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.932030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.932270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.932287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.932614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.932633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.932990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.933010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.933362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.933381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.933720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.933739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.934090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.934108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.934347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.934366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.934724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.934744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.935077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.935098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.935422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.935441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.935891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.935910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.936245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.936264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.936604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.936623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.936963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.936980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.937340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.937361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.937700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.937720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.937957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.937976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.425 [2024-11-26 19:26:58.938321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.425 [2024-11-26 19:26:58.938340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.425 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.938688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.938708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.938897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.938919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.939252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.939269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.939588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.939605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.939948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.939967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.940214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.940230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.940677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.940695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.940885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.940902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.941252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.941270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.941502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.941519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.941758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.941779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.942119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.942140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.942457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.942474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.942841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.942860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.943212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.943231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.943521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.943539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.943824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.943842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.944177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.944195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.944554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.944572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.944992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.945011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.945341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.945358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.945699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.945717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.945926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.945945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.946186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.946206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.946523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.946540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.946759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.946778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.947100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.947119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.947456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.947475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.947804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.947825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.948196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.948214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.948613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.948632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.948802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.948821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.949135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.426 [2024-11-26 19:26:58.949153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.426 qpair failed and we were unable to recover it. 00:30:46.426 [2024-11-26 19:26:58.949507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.949526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.949721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.949741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.949971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.949989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.950356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.950377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.950755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.950775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.950995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.951014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.951359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.951377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.951711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.951730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.951831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.951848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.952180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.952199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.952537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.952555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.952915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.952932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.953301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.953318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.953660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.953678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.954031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.954049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.954398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.954416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.954748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.954767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.954991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.955010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.955347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.955364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.955705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.955724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.956134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.956152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.956489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.956507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.956624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.956641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.956951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.956970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.957277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.957296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.957635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.957652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.957971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.957989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.958354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.958372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.958701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.958720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.958964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.958983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.959348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.959366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.959589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.959605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.960000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.960019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.960335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.960364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.960694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.960711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.961104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.961123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.961480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.961499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.961840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.961858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.427 qpair failed and we were unable to recover it. 00:30:46.427 [2024-11-26 19:26:58.962227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.427 [2024-11-26 19:26:58.962246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.962579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.962595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.962942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.962960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.963311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.963330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.963669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.963686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.964027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.964050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.964373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.964391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.964697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.964714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.964836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.964856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.965163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.965181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.965531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.965549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.965889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.965909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.966260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.966278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.966584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.966600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.966923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.966941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.967314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.967332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.967669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.967686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.967928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.967946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.968316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.968333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.968481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.968497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.968809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.968828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.969024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.969042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.969339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.969356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.969660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.969677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.969959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.969976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.970200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.970217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.970578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.970596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.970843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.970860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.971147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.971164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.971566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.971583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.971800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.971818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.972195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.972214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.972521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.972538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.972867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.972885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.973326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.973344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.973565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.973582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.973779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.973799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.974129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.974147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.428 qpair failed and we were unable to recover it. 00:30:46.428 [2024-11-26 19:26:58.974470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.428 [2024-11-26 19:26:58.974488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.974777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.974795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.974885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.974903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.975226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.975243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.975584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.975602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.975938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.975957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.976314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.976334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.976641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.976661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.976911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.976928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.977286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.977304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.977528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.977544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.977880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.977899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.978246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.978265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.978580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.978597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.978947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.978966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.979184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.979202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.979494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.979511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.979731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.979750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.979998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.980015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.980349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.980367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.980752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.980770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.981094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.981114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.981460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.981478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.981700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.981716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.982052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.982070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.982395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.982413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.982711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.982728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.982994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.983011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.983356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.983374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.983700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.983719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.984050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.984068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.984419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.984436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.984789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.984808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.985084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.985103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.985448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.985466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.985647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.985665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.986033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.986052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.986383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.986401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.429 [2024-11-26 19:26:58.986754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.429 [2024-11-26 19:26:58.986772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.429 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.987115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.987133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.987466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.987484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.987713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.987729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.988071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.988089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.988411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.988431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.988753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.988770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.989108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.989127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.989408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.989427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.989628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.989649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.989953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.989972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.990233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.990249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.990586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.990603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.990834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.990852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.991223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.991241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.991561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.991579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.991953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.991971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.992328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.992346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.992681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.992700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.993031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.993050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.993398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.993415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.993691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.993708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.994023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.994041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.994365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.994384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.994719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.994737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.994968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.994986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.995326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.995345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.995565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.995584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.995987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.996005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.996356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.996374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.996730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.996749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.997132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.997151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.997470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.997487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.997809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.430 [2024-11-26 19:26:58.997827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.430 qpair failed and we were unable to recover it. 00:30:46.430 [2024-11-26 19:26:58.998138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:58.998157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:58.998507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:58.998526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:58.998760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:58.998778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:58.999134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:58.999152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:58.999467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:58.999486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:58.999818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:58.999836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.000123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.000143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.000470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.000489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.000799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.000817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.001161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.001179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.001428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.001445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.001679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.001697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.002014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.002032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.002390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.002407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.002794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.002812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.003171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.003192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.003523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.003540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.003764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.003781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.004128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.004146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.004473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.004491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.004825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.004844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.005064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.005084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.005426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.005443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.005667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.005685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.006062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.006080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.006271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.006291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.006636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.006653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.006841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.006860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.007280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.007297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.007620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.007641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.007986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.008005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.008352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.008370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.008691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.008708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.009028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.009047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.009392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.009409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.009753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.009771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.010095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.010115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.010441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.010459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.431 qpair failed and we were unable to recover it. 00:30:46.431 [2024-11-26 19:26:59.010578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.431 [2024-11-26 19:26:59.010594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.011172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.011298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.011623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.011663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.012137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.012244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.012711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.012751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.013244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.013350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.013719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.013757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.014142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.014179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.014443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.014475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.014603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.014632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.014946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.014978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.015330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.015361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.015733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.015764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.016115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.016148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.432 [2024-11-26 19:26:59.016508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.432 [2024-11-26 19:26:59.016540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.432 qpair failed and we were unable to recover it. 00:30:46.705 [2024-11-26 19:26:59.016880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.016914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.017320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.017356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.017620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.017650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.018043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.018076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.018451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.018483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.018738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.018769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.019221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.019253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.019611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.019642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.020005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.020039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.020408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.020439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.020735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.020766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.021148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.021181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.021563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.021598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.022029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.022060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.022439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.022471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.022855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.022918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.023322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.023352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.023751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.023782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.023993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.024024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.024198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.024231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.024618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.024649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.025001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.025031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.025273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.025303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.025715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.025745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.026188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.026220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.026589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.026619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.027009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.027040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.027397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.027426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.027797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.027827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.028224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.028263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.028647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.028677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.029044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.029077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.029453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.029483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.029825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.029854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.030260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.030291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.030660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.030691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.031073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.031105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.031437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.031467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.031878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.031910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.032320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.032350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.032728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.032758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.033005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.033036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.033332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.033362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.033731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.033761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.034136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.034167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.034411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.034441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.034687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.034717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.035090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.035120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.035358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.035388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.035766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.706 [2024-11-26 19:26:59.035797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.706 qpair failed and we were unable to recover it. 00:30:46.706 [2024-11-26 19:26:59.036166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.036197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.036530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.036559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.036802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.036833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.037284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.037316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.037724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.037754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.038002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.038037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.038414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.038444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.038826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.038856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.039280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.039311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.039654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.039686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.039927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.039959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.040359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.040389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.040754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.040783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.041050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.041085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.041470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.041499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.041851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.041890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.042284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.042315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.042656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.042686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.043038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.043068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.043454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.043498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.043735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.043765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.044215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.044245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.044486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.044516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.044905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.044937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.045304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.045333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.045676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.045705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.046167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.046197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.046564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.046593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.046989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.047022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.047281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.047311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.047649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.047678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.047971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.048002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.048396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.048427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.048815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.048847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.049108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.049141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.049507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.049536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.049887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.049917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.050178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.050208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.050554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.050583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.050836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.050875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.051271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.051300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.051640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.051670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.051916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.051954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.052296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.707 [2024-11-26 19:26:59.052326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.707 qpair failed and we were unable to recover it. 00:30:46.707 [2024-11-26 19:26:59.052693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.052722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.052989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.053019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.053377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.053406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.053771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.053802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.054153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.054183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.054537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.054566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.054814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.054842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.055237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.055268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.055641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.055669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.055937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.055968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.056320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.056349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.056600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.056631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.056984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.057014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.057377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.057408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.057644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.057674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.058058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.058097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.058436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.058466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.058818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.058846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.059221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.059252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.059613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.059642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.059988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.060018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.060368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.060397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.060772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.060801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.061176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.061207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.061551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.061581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.061930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.061960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.062342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.062371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.062727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.062758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.063153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.063183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.063528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.063558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.063935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.063966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.064381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.064409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.064762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.064792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.065227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.065258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.065633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.065662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.065924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.065957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.066360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.066390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.066758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.066789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.067176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.067206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.067446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.067474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.067856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.067895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.068240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.068269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.068626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.068654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.069017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.069048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.069228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.069260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.069629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.069658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.708 [2024-11-26 19:26:59.070042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.708 [2024-11-26 19:26:59.070073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.708 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.070455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.070485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.070727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.070757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.071096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.071126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.071488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.071518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.071896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.071926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.072341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.072369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.072748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.072777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.073126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.073156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.073521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.073557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.073907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.073938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.074289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.074319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.074642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.074670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.075049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.075080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.075423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.075452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.075843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.075895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.076262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.076292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.076673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.076702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.077076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.077106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.077475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.077504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.077882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.077912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.078288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.078316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.078689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.078719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.079083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.079115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.079426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.079455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.079824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.079853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.080244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.080274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.080663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.080692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.080858] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1a23030 is same with the state(6) to be set 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Write completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 Read completed with error (sct=0, sc=8) 00:30:46.709 starting I/O failed 00:30:46.709 [2024-11-26 19:26:59.081222] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:46.709 [2024-11-26 19:26:59.081610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.081636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.709 [2024-11-26 19:26:59.082057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.709 [2024-11-26 19:26:59.082069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.709 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.082385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.082394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.082762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.082770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.083088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.083119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.083449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.083457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.083788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.083797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.084136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.084147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.084395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.084403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.084735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.084743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.085127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.085136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.085448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.085456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.085769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.085777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.086104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.086112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.086336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.086345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.086698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.086706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.086906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.086915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.087243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.087250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.087562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.087571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.087892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.087901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.088222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.088231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.088539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.088547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.088868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.088876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.089231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.089239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.089586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.089595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.089930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.089941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.090322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.090330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.090519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.090531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.090825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.090832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.091165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.091173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.091501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.091508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.091821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.091831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.092054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.092063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.092272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.092279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.092602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.092610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.092936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.092944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.093112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.093120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.093478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.093486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.093679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.093687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.093933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.093943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.094270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.094278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.094608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.094615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.094924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.094932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.095233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.095242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.095573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.095581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.095893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.095901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.096144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.096153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.096367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.096374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.096610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.096620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.710 [2024-11-26 19:26:59.096949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.710 [2024-11-26 19:26:59.096957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.710 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.097294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.097301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.097621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.097628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.098058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.098066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.098400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.098407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.098721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.098730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.099064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.099072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.099396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.099404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.099724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.099732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.100067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.100074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.100435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.100443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.100768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.100778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.101134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.101144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.101489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.101497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.101816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.101823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.101937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.101944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.102287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.102294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.102615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.102622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.102944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.102957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.103281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.103288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.103622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.103629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.103942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.103950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.104263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.104271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.104602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.104610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.104924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.104933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.105265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.105272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.105585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.105594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.105814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.105832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.106120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.106128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.106445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.106453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.106812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.106819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.107113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.107121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.107526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.107533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.107741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.107750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.108064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.108073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.108400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.108407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.108727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.108734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.109059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.109067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.109392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.109399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.109745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.109754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.110083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.110091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.110415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.110424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.110630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.110639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.110964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.110972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.111266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.111274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.111591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.111599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.111881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.111890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.112216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.112224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.711 [2024-11-26 19:26:59.112553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.711 [2024-11-26 19:26:59.112560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.711 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.112892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.112900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.113222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.113231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.113559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.113568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.113923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.113931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.114254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.114261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.114584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.114593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.115004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.115013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.115205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.115213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.115593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.115601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.115923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.115934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.116291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.116298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.116605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.116613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.117002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.117010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.117182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.117190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.117422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.117430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.117619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.117628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.117967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.117976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.118347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.118354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.118580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.118597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.118919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.118929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.119251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.119259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.119580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.119588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.119911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.119919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.120266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.120273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.120590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.120597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.120942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.120951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.121270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.121279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.121604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.121613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.121934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.121941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.122283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.122290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.122604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.122611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.122925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.122932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.123104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.123112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.123486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.123493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.123813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.123822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.124155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.124164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.124480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.124487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.124791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.124799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.125099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.125108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.125461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.125468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.125794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.125803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.126133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.126141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.126439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.126447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.126745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.126752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.712 [2024-11-26 19:26:59.126977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.712 [2024-11-26 19:26:59.126985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.712 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.127293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.127301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.127629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.127636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.127957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.127966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.128335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.128344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.128665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.128677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.129016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.129024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.129350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.129359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.129560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.129568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.129909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.129917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.130220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.130229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.130517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.130526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.130733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.130741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.130946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.130956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.131301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.131310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.131666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.131673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.132003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.132012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.132344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.132353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.132521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.132538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.132756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.132765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.133097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.133105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.133402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.133411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.133724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.133733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.134042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.134052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.134373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.134381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.134578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.134587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.134950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.134959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.135372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.135380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.136428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.136458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.136663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.136674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.137000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.137009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.137341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.137348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.137525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.137534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.137844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.137852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.138083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.138091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.138409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.138416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.138628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.138635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.138966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.138977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.139269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.139277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.139596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.139605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.139920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.139930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.140213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.140223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.140436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.140445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.140752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.140759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.141014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.141024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.141341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.141352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.141657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.713 [2024-11-26 19:26:59.141666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.713 qpair failed and we were unable to recover it. 00:30:46.713 [2024-11-26 19:26:59.141960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.141969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.142394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.142403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.142716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.142724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.143043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.143051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.143380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.143389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.143685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.143695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.144040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.144050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.144401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.144409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.144620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.144628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.144962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.144970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.145351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.145359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.145669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.145677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.145965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.145974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.146302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.146311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.146618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.146627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.146831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.146840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.147142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.147152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.147500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.147509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.147830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.147838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.148217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.148228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.148534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.148544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.148856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.148870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.149171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.149179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.149572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.149580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.149904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.149913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.150228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.150237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.150558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.150567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.150899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.150910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.151219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.151227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.151519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.151526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.151727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.151735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.152117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.152125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.152448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.152457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.152800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.152810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.153120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.153129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.154010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.154036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.154252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.154263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.154591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.154599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.154920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.154932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.155130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.155146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.155317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.155325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.155516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.155524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.155896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.155904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.156195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.156202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.156530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.156538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.156866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.156876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.157986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.158015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.158377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.158386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.158600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.158608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.158949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.158958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.159312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.714 [2024-11-26 19:26:59.159319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.714 qpair failed and we were unable to recover it. 00:30:46.714 [2024-11-26 19:26:59.159485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.159492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.159887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.159896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.160251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.160260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.160587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.160596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.160900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.160907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.161249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.161258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.161577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.161585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.162025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.162032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.162370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.162378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.162691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.162699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.163029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.163037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.163382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.163390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.163682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.163690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.164010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.164018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.164328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.164336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.164654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.164664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.164981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.164989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.165196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.165211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.165543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.165550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.165868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.165875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.166198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.166205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.166513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.166520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.166726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.166733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.167012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.167019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.167344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.167352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.167673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.167681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.167875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.167884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.168217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.168226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.168526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.168533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.168852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.168859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.169040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.169048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.169378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.169385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.169693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.169701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.170022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.170031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.170389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.170396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.170709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.170716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.170991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.171000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.171341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.171348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.171657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.171666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.171986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.171994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.172374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.172380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.172696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.172704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.173027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.173035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.173358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.173365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.173690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.173699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.174033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.174042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.174253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.174262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.174437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.174445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.174794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.174803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.175115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.175122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.175426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.175433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.175767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.175775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.175949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.715 [2024-11-26 19:26:59.175957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.715 qpair failed and we were unable to recover it. 00:30:46.715 [2024-11-26 19:26:59.176190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.176198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.176474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.176481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.176811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.176819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.177144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.177152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.177457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.177465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.177780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.177789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.178148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.178158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.178476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.178485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.178789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.178797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.179102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.179113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.179417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.179424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.179728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.179735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.180043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.180051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.180378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.180386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.180709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.180719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.181035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.181044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.181215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.181224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.181544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.181552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.181856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.181868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.182032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.182041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.182357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.182365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.182580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.182588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.182914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.182922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.183265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.183273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.183586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.183594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.183996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.184007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.184332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.184339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.184546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.184563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.184902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.184910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.185195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.185202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.185537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.185545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.185872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.185880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.186209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.186218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.186540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.186548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.186873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.186882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.187185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.187192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.187498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.187506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.187674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.187682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.188009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.188017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.188335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.188343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.188662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.188669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.188990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.188999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.189304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.189311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.189643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.189650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.189973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.189982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.190309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.190318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.190648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.190657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.191057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.191065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.716 [2024-11-26 19:26:59.191371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.716 [2024-11-26 19:26:59.191379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.716 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.191702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.191709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.192040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.192048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.192382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.192390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.192719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.192727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.193042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.193053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.193382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.193393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.193702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.193711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.194094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.194102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.194414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.194422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.194747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.194755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.195091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.195099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.195413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.195420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.195611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.195619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.196006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.196014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.196325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.196333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.196662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.196670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.196846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.196855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.197197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.197207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.197384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.197393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.197693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.197700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.198004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.198012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.198235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.198242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.198587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.198595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.198915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.198924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.199263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.199270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.199574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.199582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.199934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.199942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.200165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.200181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.200521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.200529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.200835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.200842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.201264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.201271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.201572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.201580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.201762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.201770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.201889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.201897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.202452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.202551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.203174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.203273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.203741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.203777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.204099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.204109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.204443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.204451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.204775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.204783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.205083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.205091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.205427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.205435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.205746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.205754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.206198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.206205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.206513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.206519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.206823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.206832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.207125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.207132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.207451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.207459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.207797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.207805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.208113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.208121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.208439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.208447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.208760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.208767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.717 [2024-11-26 19:26:59.209069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.717 [2024-11-26 19:26:59.209076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.717 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.209244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.209251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.209434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.209442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.209776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.209783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.210092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.210100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.210410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.210417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.210742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.210749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.211071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.211079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.211385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.211393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.211737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.211747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.212087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.212095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.212398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.212405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.212709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.212716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.212926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.212934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.213252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.213260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.213578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.213585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.213925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.213933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.214237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.214245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.214438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.214447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.214762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.214769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.215085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.215095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.215409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.215417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.215770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.215778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.215944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.215952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.216185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.216193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.216518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.216525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.216830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.216839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.217060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.217067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.217447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.217454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.217768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.217777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.217858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.217871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.218148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.218155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.218443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.218450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.218766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.218778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.219107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.219115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.219430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.219437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.219758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.219765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.220094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.220102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.220434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.220442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.220751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.220759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.221035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.221043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.221375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.221383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.221671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.221678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.221855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.221867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.222170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.222177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.222389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.222396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.222728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.222734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.222932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.222939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.223210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.223217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.223542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.223551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.223868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.223877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.224062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.224070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.224392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.224400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.718 [2024-11-26 19:26:59.224717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.718 [2024-11-26 19:26:59.224724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.718 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.225036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.225043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.225368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.225375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.225693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.225701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.226093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.226100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.226415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.226422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.226738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.226746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.226979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.226988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.227176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.227185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.227504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.227512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.227829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.227839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.228127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.228134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.228440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.228447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.228747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.228753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.229063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.229070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.229392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.229399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.229701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.229709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.230010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.230017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.230348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.230357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.230662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.230670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.230986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.230995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.231232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.231240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.231516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.231523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.231839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.231846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.232054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.232062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.232417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.232425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.232731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.232738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.233060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.233068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.233373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.233380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.233691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.233698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.234012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.234019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.234336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.234345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.234705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.234713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.235056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.235066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.235379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.235387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.235699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.235706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.236027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.236034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.236366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.236373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.236694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.236701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.237014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.237023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.237340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.237348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.237675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.237683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.237997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.238005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.238324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.238332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.238643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.238651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.238969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.238978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.239179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.239186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.239553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.239562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.239874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.239882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.240193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.240200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.240410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.240417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.719 qpair failed and we were unable to recover it. 00:30:46.719 [2024-11-26 19:26:59.240737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.719 [2024-11-26 19:26:59.240744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.241038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.241046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.241208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.241216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.241503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.241509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.241818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.241825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.242141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.242149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.242482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.242490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.242806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.242813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.243092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.243100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.243419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.243426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.243592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.243600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.243775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.243782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.244100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.244110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.244421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.244429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.244624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.244633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.244938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.244946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.245256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.245263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.245678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.245685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.246004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.246011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.246331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.246339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.246480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.246487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.246785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.246792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.247109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.247117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.247432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.247439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.247744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.247751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.248006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.248015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.248318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.248325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.248639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.248648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.248839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.248847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.249187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.249195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.249502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.249509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.249833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.249840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.250169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.250176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.250466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.250473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.250796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.250803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.251141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.251148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.251478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.251488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.251792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.251800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.252136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.252144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.252545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.252552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.252762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.252775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.253080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.253088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.253398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.253405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.253723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.253731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.254047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.254054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.254370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.254377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.254575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.254582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.254940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.254948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.255229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.255236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.255575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.255582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.255979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.255988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.256292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.256301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.256609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.256616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.256934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.256941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.257266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.257273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.257682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.720 [2024-11-26 19:26:59.257689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.720 qpair failed and we were unable to recover it. 00:30:46.720 [2024-11-26 19:26:59.257984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.257993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.258292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.258299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.258594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.258602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.258970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.258979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.259288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.259296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.259601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.259609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.259926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.259933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.260259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.260266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.260571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.260579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.260852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.260859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.261187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.261194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.261510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.261517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.261736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.261744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.262062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.262070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.262362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.262369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.262679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.262688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.263018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.263027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.263332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.263339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.263557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.263565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.263876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.263883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.264195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.264204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.264524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.264531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.264839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.264848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.265249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.265257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.265570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.265577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.266042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.266050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.266360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.266367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.266681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.266689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.266996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.267006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.267335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.267343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.267551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.267559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.267868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.267876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.268211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.268218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.268371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.268378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.268698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.268705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.269009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.269016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.269345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.269352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.269673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.269682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.269866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.269873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.270165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.270172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.270495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.270501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.270670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.270677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.271048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.271056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.271235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.271243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.271553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.271561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.271814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.271823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.272140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.272149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.272465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.272472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.272843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.272850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.273150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.273157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.273471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.273479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.273804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.273811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.274132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.274141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.274459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.274467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.274880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.274888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.275191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.275198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.275391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.721 [2024-11-26 19:26:59.275399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.721 qpair failed and we were unable to recover it. 00:30:46.721 [2024-11-26 19:26:59.275745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.275752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.276040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.276048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.276369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.276376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.276691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.276700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.277026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.277033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.277351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.277358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.277753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.277759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.278174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.278182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.278509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.278517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.278856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.278876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.279164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.279171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.279486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.279494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.279799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.279806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.280125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.280132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.280496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.280503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.280822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.280829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.281080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.281088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.281269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.281277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.281591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.281599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.281809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.281817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.282123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.282130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.282427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.282434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.282634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.282641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.282950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.282957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.283138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.283146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.283453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.283461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.283769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.283777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.284099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.284106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.284404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.284411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.284718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.284725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.285053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.285060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.285361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.285369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.285681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.285690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.285987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.285996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.286185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.286193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.286382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.286389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.286691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.286697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.287019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.287026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.287405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.287412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.287711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.287718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.288034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.288041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.288365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.288373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.288670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.288677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.288970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.288980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.289174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.289182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.289476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.289484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.289799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.289806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.290197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.290204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.290498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.290505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.290821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.290829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.291126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.291134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.291439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.291445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.291734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.291741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.292059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.722 [2024-11-26 19:26:59.292067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.722 qpair failed and we were unable to recover it. 00:30:46.722 [2024-11-26 19:26:59.292373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.292380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.292533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.292541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.292894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.292904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.293127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.293135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.293463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.293470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.293782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.293789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.294104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.294110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.294422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.294430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.294722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.294730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.294907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.294917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.295215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.295223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.295540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.295547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.295922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.295929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.296226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.296233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.296536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.296543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.296847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.296853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.297139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.297146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.297443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.297453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.297754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.297761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.298043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.298051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.298372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.298379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.298576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.298582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.298934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.298941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.299252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.299259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.299561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.299567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.299891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.299900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.300234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.300242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.300536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.300544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.300829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.300836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.301136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.301146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.301332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.301340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.301656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.301664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.301962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.301970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.302305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.302312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.302660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.302666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.302984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.302991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.303204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.303211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.303597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.303604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.303902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.303911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.304216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.304224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.304539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.304546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.304871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.304878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.305194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.305201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.305517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.305524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.305839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.305846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.306256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.306264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.306451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.306458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.306665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.306674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.306977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.306985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.307399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.307405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.307717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.307724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.308019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.308026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.308211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.308218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.308534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.308541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.308843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.308851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.309265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.309273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.309570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.309578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.723 qpair failed and we were unable to recover it. 00:30:46.723 [2024-11-26 19:26:59.309758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.723 [2024-11-26 19:26:59.309766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.310055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.310063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.310399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.310405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.310562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.310569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.310923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.310932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.311260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.311267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.311503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.311510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.311794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.311801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.311994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.312001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.312334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.312341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.312663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.312670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.312854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.312867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.313254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.313263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.313659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.313667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.313964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.313971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.314262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.314269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.314592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.314599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.314917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.314925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.315255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.315263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.315474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.315482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.315832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.315842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.316047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.316055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.316322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.316329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.316531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.316539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.316855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.316868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.317205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.317212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.317529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.317535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.724 [2024-11-26 19:26:59.317842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.724 [2024-11-26 19:26:59.317850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.724 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.318057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.318069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.318451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.318461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.318767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.318774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.319041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.319049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.319354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.319361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.319687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.319694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.319986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.319996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.320313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.320320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.320619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.320626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.320934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.320941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.321273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.321282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.321595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.321602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.321935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.321942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.322269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.322277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.322603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.322610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.322926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.322935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.323257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.323265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.323580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.323587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.323908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.323915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.324222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.324229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.324560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.324568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.324888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.324896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.325286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.325294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.325591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.325597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.325914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.325924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:46.999 [2024-11-26 19:26:59.326094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:46.999 [2024-11-26 19:26:59.326101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:46.999 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.326392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.326399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.326715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.326722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.327055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.327063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.327356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.327364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.327688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.327694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.328009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.328017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.328184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.328192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.328465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.328472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.328646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.328653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.329047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.329056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.329415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.329423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.329782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.329789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.330106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.330113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.330448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.330454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.330752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.330759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.331171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.331179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.331475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.331483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.331888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.331897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.332201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.332208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.332487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.332494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.332822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.332829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.333006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.333013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.333331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.333338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.333649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.333656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.333994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.334002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.334347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.334355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.334535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.334542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.334925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.334933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.335243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.335250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.335570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.335578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.335941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.335949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.336366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.336374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.336681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.336688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.337012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.337019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.337354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.337361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.337594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.337601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.337939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.337947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.000 [2024-11-26 19:26:59.338295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.000 [2024-11-26 19:26:59.338303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.000 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.338642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.338652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.338932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.338939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.339306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.339312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.339626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.339633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.339799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.339806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.340162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.340170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.340454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.340461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.340761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.340768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.340994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.341001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.341283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.341290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.341594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.341601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.341921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.341928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.342233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.342240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.342572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.342578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.342893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.342902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.343195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.343202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.343518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.343525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.343872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.343879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.344210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.344217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.344523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.344530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.344853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.344860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.345159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.345168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.345534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.345542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.345699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.345707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.345914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.345922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.346211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.346218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.346529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.346536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.346859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.346870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.347197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.347205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.347513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.347520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.347872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.347880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.348075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.348082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.348458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.348465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.348741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.348749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.349057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.349065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.349411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.349421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.349741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.349748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.350041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.001 [2024-11-26 19:26:59.350048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.001 qpair failed and we were unable to recover it. 00:30:47.001 [2024-11-26 19:26:59.350368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.350376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.350689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.350696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.351006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.351015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.351319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.351326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.351488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.351495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.351882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.351889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.352216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.352224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.352557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.352564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.352871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.352879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.353170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.353177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.353353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.353360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.353703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.353710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.354013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.354021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.354322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.354329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.354630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.354639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.354964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.354971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.355291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.355298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.355621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.355628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.356009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.356016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.356140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.356147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.356447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.356454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.356757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.356765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.357073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.357080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.357387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.357393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.357688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.357695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.357857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.357869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.358149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.358155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.358352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.358360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.358768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.358775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.359199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.359206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.359522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.359529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.359867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.359875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.360175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.360182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.360487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.360494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.360791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.360798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.361155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.361162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.361484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.361491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.361697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.002 [2024-11-26 19:26:59.361704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.002 qpair failed and we were unable to recover it. 00:30:47.002 [2024-11-26 19:26:59.362042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.362050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.362369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.362376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.362810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.362817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.363161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.363169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.363488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.363498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.363807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.363814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.364024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.364033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.364349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.364356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.364649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.364657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.364993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.365000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.365307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.365314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.365620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.365627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.365926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.365934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.366269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.366278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.366649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.366658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.366944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.366951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.367261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.367268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.367593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.367601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.367888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.367897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.368218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.368226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.368534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.368542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.368853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.368861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.369165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.369173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.369489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.369496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.369856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.369868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.370219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.370226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.370524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.370532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.370835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.370843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.371051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.371061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.371393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.371401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.371736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.371746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.003 [2024-11-26 19:26:59.372048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.003 [2024-11-26 19:26:59.372055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.003 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.372384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.372391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.372702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.372709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.373029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.373036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.373366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.373374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.373557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.373564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.373757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.373765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.374205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.374214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.374512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.374521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.374821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.374828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.375135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.375142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.375468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.375476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.375798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.375807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.376024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.376033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.376378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.376386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.376723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.376731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.376917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.376926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.377142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.377149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.377465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.377472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.377780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.377789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.378090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.378098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.378405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.378413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.378722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.378729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.379051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.379059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.379348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.379355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.379672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.379680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.379994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.380002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.380328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.380336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.380508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.380515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.380793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.380800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.381108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.381115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.381316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.381322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.381606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.381613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.381933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.381940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.382259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.382267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.382566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.382574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.382885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.382893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.383187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.383195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.004 [2024-11-26 19:26:59.383551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.004 [2024-11-26 19:26:59.383558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.004 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.383854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.383865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.384164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.384171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.384464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.384471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.384793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.384801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.385010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.385017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.385355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.385362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.385663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.385670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.385974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.385982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.386316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.386323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.386636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.386644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.386933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.386941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.387133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.387141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.387496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.387503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.387808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.387814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.388113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.388120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.388435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.388442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.388741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.388748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.388906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.388914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.389333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.389341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.389621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.389628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.389812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.389820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.390157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.390165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.390493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.390500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.390807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.390814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.391214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.391223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.391507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.391515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.391815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.391822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.392149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.392156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.392533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.392540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.392783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.392791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.393116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.393124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.393437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.393446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.393752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.393760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.394050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.394057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.394358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.394365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.394690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.394698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.395032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.395039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.395348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.005 [2024-11-26 19:26:59.395355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.005 qpair failed and we were unable to recover it. 00:30:47.005 [2024-11-26 19:26:59.395662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.395670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.395876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.395884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.396180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.396187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.396523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.396532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.396758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.396765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.396951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.396965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.397316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.397323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.397634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.397641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.397964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.397972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.398290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.398298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.398615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.398623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.398918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.398925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.399249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.399256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.399568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.399576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.399883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.399891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.400233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.400241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.400467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.400474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.400793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.400800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.401126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.401133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.401455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.401462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.401703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.401710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.401926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.401934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.402140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.402147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.402468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.402477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.402680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.402687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.403036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.403045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.403375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.403383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.403694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.403701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.403994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.404002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.404282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.404290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.404587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.404595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.404791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.404799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.405106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.405115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.405413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.405419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.405733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.405740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.405906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.405914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.406195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.406203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.406521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.006 [2024-11-26 19:26:59.406529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.006 qpair failed and we were unable to recover it. 00:30:47.006 [2024-11-26 19:26:59.406867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.406876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.407245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.407254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.407558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.407566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.407877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.407885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.408202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.408209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.408394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.408404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.408787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.408794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.409104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.409111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.409396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.409403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.409779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.409787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.410146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.410154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.410345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.410353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.410701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.410707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.411006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.411013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.411335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.411344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.411534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.411541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.411825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.411832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.412130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.412138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.412476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.412483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.412778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.412786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.413104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.413111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.413481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.413489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.413828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.413837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.414156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.414164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.414495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.414502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.414814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.414822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.415123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.415130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.415519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.415526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.415824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.415831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.416110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.416118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.416431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.416439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.416770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.416777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.417086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.417094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.417395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.417403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.417729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.417737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.418047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.418055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.418356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.418364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.418677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.418684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.007 qpair failed and we were unable to recover it. 00:30:47.007 [2024-11-26 19:26:59.419004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.007 [2024-11-26 19:26:59.419011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.419334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.419341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.419718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.419724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.419889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.419897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.420177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.420185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.420493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.420502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.420818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.420826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.421113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.421122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.421280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.421287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.421629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.421636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.421832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.421840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.422144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.422151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.422466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.422474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.422795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.422802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.423199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.423206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.423502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.423509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.423829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.423836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.424124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.424131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.424429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.424437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.424744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.424752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.425077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.425085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.425409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.425417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.425795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.425802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.426104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.426111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.426422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.426428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.426751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.426757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.426928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.426935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.427169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.427177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.427463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.427470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.427775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.427782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.428096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.428104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.428228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.428236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.428649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.428731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.429208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.429298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:47.008 qpair failed and we were unable to recover it. 00:30:47.008 [2024-11-26 19:26:59.429657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.008 [2024-11-26 19:26:59.429667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.430067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.430074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.430386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.430393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.430721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.430728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.431029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.431036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.431368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.431376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.431667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.431674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.431995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.432003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.432348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.432356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.432662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.432668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.432968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.432976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.433278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.433284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.433480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.433492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.433826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.433834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.434255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.434262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.434585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.434592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.434897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.434905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.435252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.435259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.435466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.435473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.435820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.435826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.436127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.436134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.436453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.436460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.436781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.436788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.437087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.437094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.437391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.437398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.437697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.437704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.438034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.438042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.438251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.438259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.438556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.438563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.438645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.438653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.438953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.438960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.439263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.439270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.439569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.439576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.439893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.439900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.440241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.440248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.440547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.440556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.440869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.440877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.441042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.009 [2024-11-26 19:26:59.441050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.009 qpair failed and we were unable to recover it. 00:30:47.009 [2024-11-26 19:26:59.441364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.441372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.441581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.441589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.441905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.441913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.442247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.442254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.442418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.442426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.442702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.442709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.443015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.443023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.443330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.443337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.443515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.443523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.443839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.443845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.444211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.444218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.444531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.444537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.444835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.444842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.445133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.445141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.445457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.445464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.445767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.445776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.445948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.445956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.446213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.446220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.446533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.446540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.446833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.446840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.447159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.447168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.447461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.447469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.447827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.447834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.448110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.448117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.448421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.448429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.448744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.448752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.449132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.449140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.449419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.449427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.449747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.449754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.449983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.449991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.450369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.450375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.450686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.450692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.450872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.450880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.451167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.451173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.451490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.451497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.451662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.451669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.451890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.451898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.452257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.452264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.010 qpair failed and we were unable to recover it. 00:30:47.010 [2024-11-26 19:26:59.452577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.010 [2024-11-26 19:26:59.452585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.452878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.452886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.453198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.453204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.453398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.453405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.453642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.453649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.453869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.453877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.454187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.454195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.454490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.454497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.454808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.454815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.455008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.455016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.455325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.455332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.455493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.455500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.455932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.456022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.456333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.456366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.456772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.456803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb790000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.457179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.457186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.457457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.457465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.457741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.457750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.458044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.458051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.458349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.458357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.458549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.458557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.458923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.458930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.459247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.459254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.459640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.459647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.459932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.459939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.460143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.460150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.460306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.460314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.460524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.460531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.460724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.460732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.461026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.461034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.461344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.461351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.461675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.461681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.461983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.461990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.462203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.462218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.462448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.462455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.462756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.462763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.463153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.463160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.463478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.463485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.011 [2024-11-26 19:26:59.463777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.011 [2024-11-26 19:26:59.463784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.011 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.464084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.464092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.464384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.464392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.464722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.464730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.465021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.465029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.465316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.465322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.465637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.465644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.465973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.465980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.466316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.466323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.466615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.466623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.466781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.466788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.467062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.467070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.467374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.467381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.467691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.467698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.468016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.468023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.468327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.468333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.468646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.468653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.468815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.468823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.469116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.469123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.469420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.469430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.469636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.469643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.469979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.469986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.470188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.470202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.470511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.470517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.470813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.470820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.470984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.470991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.471301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.471308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.471695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.471702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.471868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.471875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.472177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.472184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.472475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.472482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.472802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.472808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.473145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.473152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.473572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.012 [2024-11-26 19:26:59.473579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.012 qpair failed and we were unable to recover it. 00:30:47.012 [2024-11-26 19:26:59.473866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.473874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.474232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.474239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.474538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.474545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.474846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.474852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.475156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.475164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.475497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.475505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.475809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.475816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.476127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.476134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.476427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.476435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.476692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.476700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.477016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.477023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.477326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.477333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.477628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.477635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.478012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.478019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.478329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.478335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.478524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.478532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.478858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.478868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.479182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.479189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.479478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.479486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.479797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.479805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.480114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.480122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.480430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.480437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.480757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.480764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.481073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.481080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.481370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.481376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.481529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.481539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.481816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.481825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.482147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.482155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.482468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.482476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.482809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.482816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.483015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.483022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.483300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.483307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.483640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.483647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.483979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.483987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.484187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.484194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.484473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.484480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.484790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.013 [2024-11-26 19:26:59.484797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.013 qpair failed and we were unable to recover it. 00:30:47.013 [2024-11-26 19:26:59.485116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.485123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.485415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.485423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.485740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.485748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.486039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.486047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.486260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.486268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.486599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.486606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.486900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.486908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.487217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.487224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.487523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.487529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.487848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.487854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.488162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.488170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.488485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.488494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.488808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.488816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.489124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.489132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.489438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.489444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.489762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.489768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.490079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.490086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.490395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.490403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.490724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.490730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.490890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.490898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.491208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.491215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.491535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.491541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.491936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.491943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.492248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.492256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.492578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.492586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.492888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.492896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.493188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.493195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.493492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.493499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.493821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.493830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.494158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.494166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.494490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.494497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.494800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.494806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.495128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.495136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.495433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.495440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.495829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.495837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.496154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.496161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.496435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.496443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.496754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.014 [2024-11-26 19:26:59.496761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.014 qpair failed and we were unable to recover it. 00:30:47.014 [2024-11-26 19:26:59.497054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.497061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.497227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.497235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.497513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.497520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.497706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.497714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.497952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.497959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.498260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.498267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.498557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.498564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.498858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.498869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.499183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.499191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.499409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.499416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.499768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.499775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.500090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.500097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.500402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.500409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.500729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.500735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.501055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.501062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.501399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.501407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.501705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.501712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.502029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.502037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.502359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.502366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.502686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.502694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.502994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.503001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.503166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.503174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.503550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.503557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.503847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.503854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.504164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.504172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.504476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.504483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.504812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.504818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.505137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.505144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.505529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.505536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.505831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.505838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.506151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.506161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.506475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.506482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.506789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.506795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.507010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.507018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.507370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.507377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.507688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.507695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.508092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.508099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.508391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.508398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.015 qpair failed and we were unable to recover it. 00:30:47.015 [2024-11-26 19:26:59.508730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.015 [2024-11-26 19:26:59.508737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.509040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.509048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.509251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.509258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.509498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.509505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.509816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.509823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.510143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.510151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.510448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.510455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.510774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.510781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.511085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.511092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.511494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.511500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.511784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.511791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.512077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.512085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.512397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.512405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.512598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.512606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.512915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.512924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.513119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.513127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.513435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.513442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.513748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.513754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.514046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.514053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.514372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.514379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.514733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.514740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.515040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.515047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.515370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.515377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.515714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.515722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.516055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.516063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.516373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.516380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.516690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.516698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.516987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.516994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.517294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.517300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.517659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.517666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.517977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.517984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.518302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.518309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.518629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.518638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.518928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.518936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.519227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.519235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.519545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.519552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.519842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.519848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.520015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.520022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.016 [2024-11-26 19:26:59.520312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.016 [2024-11-26 19:26:59.520319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.016 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.520508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.520515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.520864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.520871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.521154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.521162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.521479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.521486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.521812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.521819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.522136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.522143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.522435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.522441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.522751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.522758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.523040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.523048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.523348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.523356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.523709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.523716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.523883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.523890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.524230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.524237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.524554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.524560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.524859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.524869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.525260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.525267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.525563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.525570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.525892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.525900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.526068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.526075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.526353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.526359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.526710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.526718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.527033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.527041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.527372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.527380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.527699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.527707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.528024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.528031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.528340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.528346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.528537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.528544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.528857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.528870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.529181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.529187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.529494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.529501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.529792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.529799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.530203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.530211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.530518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.017 [2024-11-26 19:26:59.530525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.017 qpair failed and we were unable to recover it. 00:30:47.017 [2024-11-26 19:26:59.530716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.530726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.530907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.530915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.531214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.531221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.531534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.531541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.531866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.531874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.532086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.532093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.532390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.532398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.532790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.532797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.532952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.532960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.533312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.533319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.533630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.533637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.533801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.533808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.534086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.534093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.534384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.534390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.534697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.534703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.534994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.535001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.535419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.535427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.535725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.535733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.535946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.535953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.536331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.536338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.536640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.536646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.536860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.536873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.537153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.537160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.537479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.537485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.537779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.537785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.538120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.538127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.538456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.538463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.538661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.538668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.538990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.538997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.539313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.539320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.539655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.539661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.539993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.540000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.540323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.540329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.540659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.540666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.540976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.540984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.541345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.541353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.541646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.018 [2024-11-26 19:26:59.541653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.018 qpair failed and we were unable to recover it. 00:30:47.018 [2024-11-26 19:26:59.541953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.541960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.542305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.542312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.542604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.542611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.542901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.542909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.543219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.543226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.543547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.543553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.543863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.543870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.544195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.544203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.544514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.544521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.544848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.544856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.545174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.545182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.545487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.545494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.545780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.545788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.546077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.546084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.546461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.546468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.546751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.546758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.546909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.546917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.547200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.547207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.547504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.547510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.547822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.547829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.548190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.548198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.548493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.548501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.548851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.548859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.549163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.549170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.549461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.549467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.549774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.549781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.550088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.550095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.550304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.550311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.550658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.550665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.550959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.550966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.551264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.551272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.551586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.551593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.551895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.551903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.552196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.552203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.019 [2024-11-26 19:26:59.552515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.019 [2024-11-26 19:26:59.552522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.019 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.552825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.552832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.553151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.553158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.553474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.553481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.553680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.553686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.554038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.554045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.554367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.554373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.554574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.554582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.554925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.554933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.555244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.555251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.555555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.555562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.555871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.555878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.556175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.556182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.556465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.556471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.556634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.556642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.556827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.556834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.557173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.557180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.557491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.557498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.557792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.557799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.558090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.558097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.558497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.558505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.558816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.558824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.559145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.559152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.559461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.559468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.559757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.559764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.559977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.559985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.560337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.560343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.560551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.560558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.560878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.560885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.561208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.561215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.561401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.561408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.020 [2024-11-26 19:26:59.561702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.020 [2024-11-26 19:26:59.561708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.020 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.562025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.562032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.562354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.562361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.562694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.562700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.563009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.563017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.563340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.563348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.563649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.563656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.563936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.563943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.564229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.564236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.564621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.564627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.564939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.564946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.565275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.565282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.565579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.565586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.565909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.565916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.566076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.566083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.566307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.566314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.566692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.566700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.567013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.567020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.567313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.567320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.567630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.567638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.567881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.567888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.568184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.568191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.568472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.568479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.568793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.568799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.569149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.569157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.569467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.569474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.569805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.569812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.570098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.570105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.570414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.570422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.570693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.570700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.570874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.570881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.571250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.571256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.571444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.571451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.571768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.571774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.572088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.021 [2024-11-26 19:26:59.572095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.021 qpair failed and we were unable to recover it. 00:30:47.021 [2024-11-26 19:26:59.572392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.572399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.572690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.572697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.573047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.573054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.573341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.573348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.573554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.573561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.573774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.573781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.574052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.574059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.574360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.574367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.574546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.574554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.574926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.574932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.575256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.575264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.575571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.575578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.575774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.575781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.575976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.575983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.576094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.576100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.576407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.576414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.576725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.576731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.577070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.577077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.577399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.577405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.577726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.577732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.578137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.578144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.578433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.578440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.578720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.578727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.578987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.578995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.579295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.579303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.579455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.579463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.579738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.579745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.580041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.580048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.580319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.580325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.580641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.580648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.580978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.580985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.581153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.581162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.581444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.581450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.581738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.581745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.582038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.022 [2024-11-26 19:26:59.582045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.022 qpair failed and we were unable to recover it. 00:30:47.022 [2024-11-26 19:26:59.582214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.582221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.582504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.582510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.582835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.582841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.583152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.583160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.583471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.583478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.583655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.583662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.583943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.583951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.584233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.584239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.584618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.584625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.584916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.584923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.585112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.585119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.585391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.585398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.585690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.585697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.586016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.586022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.586342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.586348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.586642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.586650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.586960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.586967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.587289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.587296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.587590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.587596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.587901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.587908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.588309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.588316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.588507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.588514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.588833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.588840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.589085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.589093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.589396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.589403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.589690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.589696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.589853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.589860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.590246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.590253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.590564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.590571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.590879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.590887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.591187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.591194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.591470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.591476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.591788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.591795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.592183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.592190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.592480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.023 [2024-11-26 19:26:59.592486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.023 qpair failed and we were unable to recover it. 00:30:47.023 [2024-11-26 19:26:59.592682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.592689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.593010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.593018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.593334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.593341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.593631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.593637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.593925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.593932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.594240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.594247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.594561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.594567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.594863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.594871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.595188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.595195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.595356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.595363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.595657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.595663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.596061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.596067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.596366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.596373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.596761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.596769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.596949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.596957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.597266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.597273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.597604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.597611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.597899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.597906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.598212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.598219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.598551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.598558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.598872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.598880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.599168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.599175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.599487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.599494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.599801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.599809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.600115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.600121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.600279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.600287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.600705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.600711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.600871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.600879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.601174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.601181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.601498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.601505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.601823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.601830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.602120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.602127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.602444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.602451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.602750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.602758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.603040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.603047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.603343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.603351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.603660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.024 [2024-11-26 19:26:59.603667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.024 qpair failed and we were unable to recover it. 00:30:47.024 [2024-11-26 19:26:59.603974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.603981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.604301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.604308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.604626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.604632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.604916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.604923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.605230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.605238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.605546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.605553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.605866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.605873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.606196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.606203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.606496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.606504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.606832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.606839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.607126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.607133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.607407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.607414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.025 [2024-11-26 19:26:59.607723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.025 [2024-11-26 19:26:59.607729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.025 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.608051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.608060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.608374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.608381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.608673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.608681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.609004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.609011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.609322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.609330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.609616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.609623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.609923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.609930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.610255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.610261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.610585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.610592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.610895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.610903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.611224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.611233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.611530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.611537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.611890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.611897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.612183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.612190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.612501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.612508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.612820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.612827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.613007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.613014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.613168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.613175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.613335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.613342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.613618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.613625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.613831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.613837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.614146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.614153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.614424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.614432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.614756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.614763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.615074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.615081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.615249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.310 [2024-11-26 19:26:59.615256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.310 qpair failed and we were unable to recover it. 00:30:47.310 [2024-11-26 19:26:59.615425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.615431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.615734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.615741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.616158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.616166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.616459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.616466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.616750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.616757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.617065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.617072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.617370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.617377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.617703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.617709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.618001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.618008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.618192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.618199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.618540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.618546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.618848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.618855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.619162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.619170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.619507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.619514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.619806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.619812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.620128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.620136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.620307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.620315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.620598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.620605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.620898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.620905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.621212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.621219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.621513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.621521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh: line 36: 4064757 Killed "${NVMF_APP[@]}" "$@" 00:30:47.311 [2024-11-26 19:26:59.621738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.621745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.622111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.622118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.622274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.622281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@48 -- # disconnect_init 10.0.0.2 00:30:47.311 [2024-11-26 19:26:59.622688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.622696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:30:47.311 [2024-11-26 19:26:59.623004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.623011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:30:47.311 [2024-11-26 19:26:59.623305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.623312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:30:47.311 [2024-11-26 19:26:59.623626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.623633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:47.311 [2024-11-26 19:26:59.623854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.623864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.624213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.624220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.624497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.624504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.624705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.624711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.624886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.624893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.625098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.625106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.625599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.625637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1a26490 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.311 [2024-11-26 19:26:59.626090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.311 [2024-11-26 19:26:59.626129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1a26490 with addr=10.0.0.2, port=4420 00:30:47.311 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.626423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.626432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.626746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.626754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.626975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.626982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.627328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.627334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.627650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.627658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.627867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.627875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.628210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.628217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.628535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.628542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.628866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.628873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.629155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.629163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.629338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.629346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.629678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.629686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.629865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.629875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.630193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.630201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.630512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.630519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.630829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.630836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.631145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.631153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # nvmfpid=4065791 00:30:47.312 [2024-11-26 19:26:59.631462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.631470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.631637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.631645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # waitforlisten 4065791 00:30:47.312 [2024-11-26 19:26:59.631979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.631987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # '[' -z 4065791 ']' 00:30:47.312 [2024-11-26 19:26:59.632161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.632170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.632377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.632384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:47.312 [2024-11-26 19:26:59.632710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.632717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:47.312 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:47.312 [2024-11-26 19:26:59.633061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.633070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:47.312 [2024-11-26 19:26:59.633279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.633287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 19:26:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:47.312 [2024-11-26 19:26:59.633601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.633610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.633933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.633941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.634124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.634132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.634433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.634441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.634652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.634660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.634894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.634902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.635221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.635229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.635562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.635570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.312 qpair failed and we were unable to recover it. 00:30:47.312 [2024-11-26 19:26:59.635772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.312 [2024-11-26 19:26:59.635779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.636051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.636061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.636242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.636249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.636480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.636488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.636787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.636795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.637041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.637049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.637250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.637258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.637537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.637545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.637732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.637739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.638044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.638052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.638235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.638243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.638568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.638575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.638880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.638888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.639202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.639209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.639416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.639424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.639651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.639659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.639986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.639995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.640180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.640189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.640342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.640350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.640539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.640547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.640871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.640879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.641076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.641084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.641393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.641401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.641611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.641619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.641803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.641810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.642072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.642080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.642355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.642363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.642543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.642550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.642849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.642856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.643208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.643216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.643393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.643401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.643769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.643776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.644130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.644139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.644431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.644439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.644785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.644793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.645136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.645144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.645353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.645360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.645680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.645687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.313 [2024-11-26 19:26:59.645999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.313 [2024-11-26 19:26:59.646006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.313 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.646349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.646356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.646659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.646665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.646720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.646729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.646909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.646916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.647231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.647239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.647561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.647567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.647880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.647887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.648065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.648073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.648485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.648491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.648778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.648785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.649068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.649076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.649404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.649410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.649716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.649723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.650029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.650037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.650363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.650370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.650653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.650661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.650986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.650993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.651314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.651321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.651519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.651526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.651738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.651745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.652047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.652054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.652348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.652355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.652549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.652556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.652871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.652878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.653169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.653176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.653476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.653483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.653783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.653790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.654113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.654120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.654572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.654579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.654870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.654878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.655230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.655237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.655404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.655410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.314 [2024-11-26 19:26:59.655607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.314 [2024-11-26 19:26:59.655613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.314 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.655986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.655992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.656296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.656304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.656519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.656526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.656859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.656872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.657017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.657024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.657379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.657386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.657677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.657684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.658002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.658009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.658205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.658213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.658414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.658422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.658713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.658720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.659038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.659046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.659232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.659240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.659425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.659432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.659755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.659763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.660096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.660104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.660424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.660430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.660756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.660762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.661082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.661089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.661395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.661402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.661598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.661605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.661770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.661776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.661941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.661949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.662312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.662319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.662621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.662628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.662831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.662838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.663025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.663032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.663335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.663342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.663617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.663624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.663923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.663930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.664228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.664235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.664540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.664548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.664871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.664879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.665166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.665172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.665465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.665471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.665562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.665569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.665768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.665775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.315 [2024-11-26 19:26:59.666089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.315 [2024-11-26 19:26:59.666096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.315 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.666421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.666428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.666754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.666760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.667092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.667098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.667423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.667430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.667768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.667776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.667811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.667819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.668123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.668131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.668294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.668301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.668660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.668667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.668877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.668890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.669227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.669233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.669530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.669539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.669852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.669859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.670045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.670052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.670336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.670342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.670661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.670669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.670986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.670992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.671307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.671314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.671653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.671660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.671992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.672000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.672178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.672186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.672479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.672486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.672699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.672706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.673012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.673020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.673397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.673404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.673590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.673597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.673879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.673886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.674090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.674098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.674419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.674426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.674653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.674660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.674985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.674992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.675286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.675292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.675469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.675476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.675713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.675720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.676060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.676068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.676409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.676416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.676743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.676750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.677082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.316 [2024-11-26 19:26:59.677090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.316 qpair failed and we were unable to recover it. 00:30:47.316 [2024-11-26 19:26:59.677331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.677338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.677517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.677524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.677665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.677672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.677973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.677981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.678362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.678370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.678675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.678683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.678850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.678857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.679232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.679239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.679435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.679442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.679815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.679822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.680125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.680132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.680319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.680326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.680610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.680617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.680763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.680772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.681074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.681155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.681412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.681445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb784000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.681761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.681770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.682027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.682035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.682383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.682390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.682709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.682716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.683030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.683037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.683368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.683375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.683721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.683729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.684034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.684041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.684342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.684349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.684669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.684677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.684976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.684984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.685191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.685198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.685392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.685399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.685701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.685708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.685839] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:30:47.317 [2024-11-26 19:26:59.685895] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:47.317 [2024-11-26 19:26:59.685975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.685982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.686342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.686349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.686672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.686678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.686876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.686883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.687136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.687145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.687325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.687334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.687653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.317 [2024-11-26 19:26:59.687662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.317 qpair failed and we were unable to recover it. 00:30:47.317 [2024-11-26 19:26:59.687848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.687857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.688159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.688168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.688532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.688540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.688950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.688959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.689272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.689280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.689608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.689617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.689932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.689941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.690276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.690284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.690590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.690598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.690777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.690787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.690997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.691006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.691384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.691393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.691705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.691714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.691941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.691950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.692240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.692249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.692580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.692592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.692780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.692787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.693152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.693159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.693457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.693464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.693783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.693790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.693969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.693977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.694303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.694310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.694614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.694621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.694846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.694854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.695187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.695194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.695499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.695506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.695805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.695812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.696097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.696105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.696453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.696460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.696763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.696771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.697081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.697088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.697322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.697329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.697629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.697636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.697937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.697945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.698243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.698250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.698578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.698585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.698891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.698898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.699080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.318 [2024-11-26 19:26:59.699088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.318 qpair failed and we were unable to recover it. 00:30:47.318 [2024-11-26 19:26:59.699296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.699303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.699602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.699609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.699912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.699919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.700219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.700226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.700552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.700559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.700714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.700721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.701113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.701120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.701284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.701299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.701542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.701550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.701849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.701855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.702166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.702173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.702485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.702492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.702818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.702825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.703130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.703137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.703457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.703465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.703513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.703521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.703832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.703839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.704121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.704130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.704458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.704465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.704800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.704807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.705143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.705150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.705477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.705484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.705705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.705712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.706057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.706064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.706419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.706426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.706743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.706750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.707071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.707079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.707390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.707397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.707716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.707722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.707941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.707956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.708176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.708182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.708504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.708511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.708803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.708810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.709009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.319 [2024-11-26 19:26:59.709023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.319 qpair failed and we were unable to recover it. 00:30:47.319 [2024-11-26 19:26:59.709357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.709363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.709518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.709525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.709794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.709800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.709979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.709987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.710274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.710281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.710420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.710427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.710700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.710707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.711009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.711017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.711340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.711347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.711530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.711538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.711837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.711845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.712181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.712188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.712481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.712488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.712680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.712686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.713034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.713041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.713355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.713362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.713677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.713683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.714013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.714021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.714358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.714366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.714557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.714563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.714877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.714885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.715227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.715233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.715557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.715564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.715900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.715910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.716299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.716306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.716594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.716601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.716892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.716899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.717209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.717216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.717413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.717420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.717781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.717787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.717962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.717971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.718265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.718273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.718543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.718551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.718605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.718611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.718928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.718936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.719319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.719326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.719633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.719640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.720042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.320 [2024-11-26 19:26:59.720050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.320 qpair failed and we were unable to recover it. 00:30:47.320 [2024-11-26 19:26:59.720231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.720238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.720465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.720472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.720794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.720801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.721105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.721112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.721492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.721500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.721831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.721838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.722146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.722153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.722486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.722494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.722702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.722709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.722996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.723003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.723217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.723224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.723520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.723526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.723850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.723857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.724196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.724203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.724388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.724395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.724760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.724768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.724845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.724851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.725176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.725183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.725486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.725493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.725821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.725828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.726161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.726168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.726538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.726545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.726891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.726899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.727355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.727363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.727678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.727685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.727981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.727991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.728330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.728338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.728643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.728649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.728978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.728985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.729288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.729295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.729499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.729507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.729577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.729583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.729893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.729901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.730120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.730127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.730420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.730427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.730715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.730722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.731067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.731074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.731429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.321 [2024-11-26 19:26:59.731436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.321 qpair failed and we were unable to recover it. 00:30:47.321 [2024-11-26 19:26:59.731601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.731608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.731978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.731985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.732283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.732290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.732607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.732614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.732921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.732928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.733300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.733307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.733597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.733604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.733784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.733791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.733990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.733997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.734337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.734343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.734553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.734559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.734771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.734777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.734968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.734978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.735282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.735289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.735502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.735510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.735855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.735863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.736205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.736212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.736477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.736483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.736814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.736820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.737129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.737136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.737462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.737469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.737765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.737772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.738060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.738068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.738393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.738401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.738606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.738613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.738777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.738785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.739158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.739166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.739293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.739302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.739599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.739607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.739995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.740003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.740256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.740263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.740693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.740699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.741051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.741059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.741113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.741121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.741403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.741410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.741554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.741560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.741848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.741854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.322 [2024-11-26 19:26:59.742223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.322 [2024-11-26 19:26:59.742230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.322 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.742574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.742581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.742875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.742882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.743204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.743211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.743413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.743420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.743638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.743645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.743984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.743991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.744194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.744201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.744517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.744524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.744811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.744818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.745162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.745169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.745484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.745490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.745814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.745821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.745986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.745994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.746268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.746275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.746577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.746584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.746751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.746758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.747000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.747007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.747312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.747320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.747650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.747658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.747837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.747845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.748184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.748192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.748492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.748500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.748587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.748594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.748864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.748871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.749236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.749243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.749433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.749440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.749778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.749785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.750144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.750151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.750487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.750493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.750806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.750814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.751133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.751140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.751469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.751476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.751771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.751779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.323 [2024-11-26 19:26:59.751971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.323 [2024-11-26 19:26:59.751979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.323 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.752284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.752291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.752583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.752591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.752793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.752799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.753073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.753080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.753280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.753287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.753647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.753654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.753973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.753980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.754279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.754286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.754669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.754676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.754972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.754979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.755273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.755280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.755625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.755632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.755676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.755684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.755968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.755976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.756322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.756328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.756706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.756713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.757034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.757041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.757221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.757229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.757565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.757571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.757876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.757883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.758177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.758184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.758377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.758384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.758596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.758603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.758909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.758917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.759262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.759269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.759580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.759586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.759901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.759908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.760135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.760142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.760475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.760481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.760662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.760668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.760867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.760873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.761199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.761206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.761547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.761554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.761769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.761776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.762163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.762171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.762219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.762227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.762565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.762572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.324 [2024-11-26 19:26:59.762876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.324 [2024-11-26 19:26:59.762883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.324 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.763173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.763180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.763518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.763525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.763850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.763857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.764163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.764170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.764498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.764505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.764810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.764817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.765188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.765196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.765510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.765517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.765714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.765721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.766034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.766041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.766344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.766350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.766771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.766778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.766934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.766941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.767307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.767314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.767608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.767615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.767911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.767918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.768099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.768108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.768416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.768422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.768738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.768745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.769059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.769066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.769453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.769460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.769508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.769515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.769844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.769850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.770221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.770228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.770521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.770529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.770819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.770826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.771143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.771151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.771355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.771361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.771711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.771718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.772016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.772023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.772067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.772073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.772425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.772432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.772702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.772709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.772971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.772978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.773344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.773351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.773522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.773530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.773703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.773709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.773914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.325 [2024-11-26 19:26:59.773923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.325 qpair failed and we were unable to recover it. 00:30:47.325 [2024-11-26 19:26:59.774248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.774255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.774474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.774480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.774840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.774847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.775211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.775218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.775499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.775506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.775663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.775670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.776053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.776060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.776461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.776468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.776767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.776773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.776963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.776971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.777304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.777311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.777605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.777612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.777931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.777938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.778252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.778258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.778559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.778565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.778884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.778891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.779072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.779079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.779366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.779374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.779697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.779704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.779999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.780006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.780391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.780398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.780710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.780717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.781036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.781044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.781381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.781387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.781683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.781690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.782006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.782013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.782341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.782348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.782671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.782678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.783007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.783014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.783334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.783341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.783752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.783760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.784049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.784056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.784312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.784319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.784728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.784735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.785113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.785120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.785468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.785474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.785746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.785753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.785970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.326 [2024-11-26 19:26:59.785977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.326 qpair failed and we were unable to recover it. 00:30:47.326 [2024-11-26 19:26:59.786242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.786248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.786551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.786559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.786721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.786729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.787011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.787018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.787316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.787322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.787520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.787535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.787857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.787867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.788183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.788190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.788512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.788519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.788842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.788848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.789158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.789165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.789487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.789493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.789674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.789681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.790043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.790050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.790233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.790241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.790535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.790542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.790843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.790849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.791035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.791042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.791347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.791354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.791638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.791644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.791826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.791833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.792207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.792214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.792510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.792517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.792839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.792846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.793028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.793035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.793203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.793209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.793293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.793300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.793585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.793593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.793890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.793898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.794234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.794241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.794276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.794283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.794363] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:30:47.327 [2024-11-26 19:26:59.794481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.794488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.794658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.794665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.794960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.794968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.795372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.795379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.795691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.795698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.795910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.795918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.796119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.327 [2024-11-26 19:26:59.796127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.327 qpair failed and we were unable to recover it. 00:30:47.327 [2024-11-26 19:26:59.796428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.796435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.796756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.796763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.797087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.797094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.797260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.797270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.797671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.797678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.798001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.798008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.798308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.798316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.798630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.798638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.798890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.798899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.799086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.799092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.799358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.799366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.799556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.799563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.799918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.799926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.800009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.800016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.800312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.800319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.800619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.800626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.800924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.800931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.801300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.801308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.801479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.801487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.801757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.801764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.802082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.802089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.802380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.802387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.802719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.802727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.803038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.803045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.803261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.803269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.803459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.803465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.803736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.803744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.804065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.804072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.804391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.804398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.804602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.804609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.804932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.804939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.805274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.805281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.805598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.328 [2024-11-26 19:26:59.805605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.328 qpair failed and we were unable to recover it. 00:30:47.328 [2024-11-26 19:26:59.805923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.805930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.806247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.806255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.806444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.806452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.806730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.806736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.807032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.807039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.807213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.807221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.807543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.807550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.807842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.807848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.808039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.808048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.808297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.808304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.808610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.808619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.808930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.808937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.809255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.809262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.809554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.809561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.809826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.809834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.810189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.810197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.810491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.810499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.810806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.810813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.811166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.811173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.811490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.811497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.811825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.811832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.812152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.812159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.812447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.812454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.812747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.812754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.813050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.813058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.813380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.813387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.813699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.813706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.814010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.814018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.814347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.814355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.814665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.814673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.814956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.814963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.815260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.815267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.815479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.815485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.815591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.815597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.815928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.815936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.816313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.816320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.816603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.816610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.816910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.816918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.329 qpair failed and we were unable to recover it. 00:30:47.329 [2024-11-26 19:26:59.817292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.329 [2024-11-26 19:26:59.817300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.817605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.817613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.817890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.817898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.818235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.818242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.818554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.818560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.818859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.818869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.819162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.819168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.819527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.819534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.819816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.819823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.820126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.820134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.820448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.820455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.820780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.820788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.820967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.820978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.821306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.821314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.821665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.821673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.821952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.821960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.822258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.822265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.822509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.822516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.822822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.822829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.823122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.823130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.823422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.823428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.823727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.823734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.824058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.824065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.824386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.824393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.824711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.824718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.825033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.825041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.825269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.825276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.825547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.825555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.825872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.825880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.826048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.826056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.826365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.826372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.826681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.826688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.826983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.826990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.827289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.827297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.827465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.827474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.827787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.827794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.828089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.828096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.828429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.828436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.330 [2024-11-26 19:26:59.828764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.330 [2024-11-26 19:26:59.828772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.330 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.829123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.829132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.829450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.829457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.829774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.829780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.829866] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:47.331 [2024-11-26 19:26:59.829893] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:47.331 [2024-11-26 19:26:59.829901] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:47.331 [2024-11-26 19:26:59.829907] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:47.331 [2024-11-26 19:26:59.829913] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:47.331 [2024-11-26 19:26:59.830081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.830089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.830386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.830393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.830795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.830801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.831100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.831108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.831299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.831306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.831476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.831485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.831499] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:30:47.331 [2024-11-26 19:26:59.831660] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:30:47.331 [2024-11-26 19:26:59.831763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.831775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.831945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.831954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.831916] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:30:47.331 [2024-11-26 19:26:59.831917] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:30:47.331 [2024-11-26 19:26:59.832287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.832295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.832608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.832615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.832819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.832828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.833179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.833186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.833494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.833501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.833837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.833844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.834193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.834201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.834532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.834539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.834854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.834864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.835044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.835051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.835426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.835433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.835745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.835752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.836090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.836100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.836298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.836305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.836629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.836636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.836938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.836945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.837151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.837157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.837517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.837525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.837733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.837741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.837926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.837933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.838306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.838313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.838613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.838620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.838933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.838941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.331 [2024-11-26 19:26:59.839327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.331 [2024-11-26 19:26:59.839335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.331 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.839649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.839657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.839967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.839974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.840290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.840297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.840624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.840630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.840991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.840998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.841192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.841199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.841454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.841461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.841663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.841669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.841917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.841924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.842211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.842217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.842558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.842565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.842870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.842878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.843211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.843218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.843552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.843559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.843906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.843913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.844223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.844230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.844526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.844534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.844709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.844717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.845037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.845045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.845251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.845258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.845430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.845437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.845669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.845677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.846025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.846033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.846229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.846236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.846477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.846484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.846804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.846810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.847245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.847253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.847454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.847462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.847775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.847798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.847991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.847999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.848287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.848294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.848457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.848463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.848736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.848743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.848950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.848957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.849164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.332 [2024-11-26 19:26:59.849171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.332 qpair failed and we were unable to recover it. 00:30:47.332 [2024-11-26 19:26:59.849496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.849503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.849827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.849834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.850049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.850057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.850349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.850357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.850664] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.850673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.850986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.850994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.851181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.851193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.851359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.851366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.851547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.851555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.851874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.851882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.852031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.852037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.852351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.852359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.852669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.852677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.852989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.852997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.853309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.853316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.853649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.853656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.853868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.853876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.854080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.854086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.854290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.854297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.854596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.854603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.854927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.854935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.855301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.855308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.855469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.855477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.855656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.855665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.855986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.855993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.856164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.856171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.856454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.856462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.856726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.856734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.857035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.857042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.857231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.857239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.857426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.857433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.857732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.857739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.858031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.858039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.858325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.858334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.858665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.858672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.859039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.859046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.333 [2024-11-26 19:26:59.859344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.333 [2024-11-26 19:26:59.859351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.333 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.859516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.859523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.859817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.859825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.860140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.860147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.860441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.860448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.860765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.860772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.861093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.861100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.861282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.861289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.861350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.861357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.861531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.861538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.861584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.861591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.861875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.861883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.862165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.862172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.862487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.862494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.862806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.862814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.862978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.862986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.863181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.863188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.863503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.863511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.863691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.863699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.863899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.863907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.864201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.864209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.864515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.864523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.864683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.864692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.864999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.865007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.865369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.865376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.865705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.865714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.866045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.866053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.866446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.866455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.866616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.866623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.866959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.866967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.867296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.867304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.867615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.867624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.867938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.867946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.868144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.868151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.868474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.868482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.868699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.868707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.868858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.868869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.868911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.868921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.869291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.334 [2024-11-26 19:26:59.869298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.334 qpair failed and we were unable to recover it. 00:30:47.334 [2024-11-26 19:26:59.869474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.869481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.869774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.869781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.869944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.869952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.870195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.870203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.870504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.870511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.870822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.870829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.871001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.871008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.871229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.871236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.871408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.871415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.871578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.871585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.871893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.871902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.872215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.872222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.872546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.872552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.872788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.872795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.872967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.872976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.873180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.873187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.873347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.873354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.873557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.873565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.873931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.873939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.874251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.874258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.874441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.874448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.874615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.874621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.874931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.874938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.875120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.875128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.875428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.875435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.875781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.875790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.875950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.875958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.876326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.876332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.876632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.876639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.876830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.876838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.877158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.877165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.877460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.877467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.877786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.877793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.878121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.878128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.878433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.878440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.878826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.878834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.879002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.879009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.879287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.335 [2024-11-26 19:26:59.879294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.335 qpair failed and we were unable to recover it. 00:30:47.335 [2024-11-26 19:26:59.879488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.879495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.879834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.879841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.880018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.880025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.880330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.880336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.880661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.880668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.880985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.880992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.881081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.881087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.881250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.881258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.881637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.881645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.881956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.881963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.882287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.882294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.882328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.882335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.882655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.882661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.882853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.882860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.883117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.883125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.883424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.883431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.883600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.883608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.883888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.883895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.884154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.884160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.884460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.884466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.884753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.884759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.884949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.884956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.885247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.885254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.885305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.885311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.885606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.885612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.885910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.885917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.886225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.886233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.886422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.886431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.886611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.886618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.886928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.886935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.887325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.887331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.887662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.887668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.887708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.887715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.887940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.887947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.888259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.888265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.888580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.888586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.888754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.888761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.889135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.889142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.336 qpair failed and we were unable to recover it. 00:30:47.336 [2024-11-26 19:26:59.889459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.336 [2024-11-26 19:26:59.889466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.889663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.889670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.890055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.890061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.890354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.890360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.890664] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.890670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.890985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.890992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.891258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.891265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.891623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.891630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.891954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.891961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.892136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.892143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.892190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.892196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.892416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.892422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.892750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.892757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.893101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.893109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.893154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.893161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.893457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.893464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.893660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.893668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.894018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.894025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.894344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.894351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.894531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.894539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.894718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.894724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.895078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.895085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.895261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.895268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.895546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.895552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.895850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.895856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.896201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.896208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.896503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.896509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.896721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.896728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.896912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.896918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.897221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.897230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.897534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.897540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.897864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.337 [2024-11-26 19:26:59.897871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.337 qpair failed and we were unable to recover it. 00:30:47.337 [2024-11-26 19:26:59.898205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.898212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.898512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.898520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.898676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.898684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.898969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.898977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.899297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.899305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.899474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.899481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.899792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.899798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.899971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.899979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.900205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.900211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.900403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.900410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.900768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.900775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.901098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.901105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.901394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.901401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.901592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.901598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.901634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.901642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.901945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.901951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.902133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.902141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.902476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.902483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.902772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.902779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.902815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.902822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.903147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.903154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.903444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.903450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.903755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.903762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.903950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.903956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.904203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.904210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.904452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.904459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.904666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.904673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.904921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.904929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.905085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.905091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.905248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.905254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.905419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.905426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.905610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.905616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.905924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.905931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.906241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.906247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.906570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.906576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.906908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.906915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.907277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.907283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.907598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.338 [2024-11-26 19:26:59.907606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.338 qpair failed and we were unable to recover it. 00:30:47.338 [2024-11-26 19:26:59.907772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.339 [2024-11-26 19:26:59.907779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.339 qpair failed and we were unable to recover it. 00:30:47.339 [2024-11-26 19:26:59.908064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.339 [2024-11-26 19:26:59.908072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.339 qpair failed and we were unable to recover it. 00:30:47.339 [2024-11-26 19:26:59.908400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.339 [2024-11-26 19:26:59.908407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.339 qpair failed and we were unable to recover it. 00:30:47.339 [2024-11-26 19:26:59.908722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.339 [2024-11-26 19:26:59.908729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.339 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.909049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.909058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.909264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.909273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.909639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.909646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.910014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.910021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.910192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.910200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.910394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.910401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.910591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.910599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.910928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.910935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.911114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.911120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.911573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.911580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.911868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.911876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.912158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.912165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.912485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.912493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.912797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.912804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.913111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.913119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.913424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.913431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.913748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.913755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.914039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.914046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.914378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.914385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.914545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.625 [2024-11-26 19:26:59.914552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.625 qpair failed and we were unable to recover it. 00:30:47.625 [2024-11-26 19:26:59.914751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.914757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.915124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.915131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.915527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.915533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.915851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.915858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.915940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.915947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.916285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.916291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.916463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.916471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.916750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.916757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.917085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.917091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.917267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.917274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.917557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.917564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.917866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.917873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.918204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.918211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.918514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.918521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.918699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.918706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.918920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.918929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.919083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.919090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.919381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.919388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.919779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.919786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.920089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.920096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.920398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.920406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.920605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.920612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.920943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.920951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.921236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.921243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.921545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.921551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.921842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.921848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.922157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.922164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.922485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.922492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.922778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.922784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.923123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.923130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.923291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.923299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.923586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.923594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.923912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.923920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.924251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.924258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.924580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.924586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.924774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.924780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.924930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.924937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.925121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.925128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.626 qpair failed and we were unable to recover it. 00:30:47.626 [2024-11-26 19:26:59.925163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.626 [2024-11-26 19:26:59.925170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.925540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.925547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.925838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.925844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.926134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.926142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.926467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.926473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.926806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.926813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.927197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.927204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.927418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.927425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.927737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.927743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.927903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.927910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.928077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.928084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.928296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.928304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.928479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.928486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.928668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.928675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.928883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.928889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.929176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.929183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.929489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.929495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.929702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.929710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.929878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.929885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.930185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.930192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.930525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.930532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.930889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.930896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.931161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.931168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.931331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.931337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.931613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.931620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.931657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.931663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.931964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.931971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.932288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.932295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.932589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.932595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.932900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.932907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.933242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.933248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.933545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.933551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.933896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.933904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.934123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.934129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.934299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.934306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.934599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.934606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.934841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.934848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.935278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.935285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.627 [2024-11-26 19:26:59.935481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.627 [2024-11-26 19:26:59.935488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.627 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.935802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.935808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.936132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.936145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.936317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.936324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.936492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.936499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.936872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.936879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.937096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.937104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.937399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.937406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.937566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.937573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.937801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.937809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.938087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.938095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.938252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.938258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.938449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.938455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.938865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.938872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.939188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.939194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.939511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.939518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.939834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.939841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.939881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.939888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.940253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.940261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.940452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.940461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.940612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.940619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.940659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.940666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.940908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.940915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.941080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.941087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.941255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.941262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.941577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.941584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.941879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.941886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.942067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.942074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.942244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.942250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.942545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.942551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.942888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.942895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.943196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.943203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.943525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.943533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.943705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.943713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.944031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.944039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.944353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.944360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.944526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.944534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.944811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.944818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.628 [2024-11-26 19:26:59.944997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.628 [2024-11-26 19:26:59.945005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.628 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.945315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.945322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.945635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.945641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.946023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.946030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.946201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.946209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.946538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.946544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.946920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.946927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.947238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.947245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.947436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.947444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.947488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.947495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.947795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.947801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.947959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.947974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.948313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.948320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.948668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.948675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.948879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.948886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.949071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.949077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.949376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.949382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.949680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.949686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.949884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.949891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.950132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.950139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.950304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.950312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.950643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.950651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.950943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.950950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.951175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.951182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.951460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.951467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.951781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.951787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.951950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.951958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.952244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.952251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.952478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.952485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.952723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.952730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.952925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.952933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.953089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.953096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.953383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.953390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.953725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.953732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.629 [2024-11-26 19:26:59.954069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.629 [2024-11-26 19:26:59.954076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.629 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.954270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.954277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.954349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.954356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.954647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.954654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.954987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.954994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.955287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.955294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.955454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.955461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.955806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.955812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.956161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.956168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.956461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.956468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.956788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.956795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.956965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.956972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.957246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.957252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.957564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.957570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.957964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.957972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.958269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.958276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.958357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.958364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.958633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.958640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.958959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.958966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.959261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.959268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.959541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.959548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.959771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.959777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.960058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.960065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.960369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.960376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.960673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.960679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.960844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.960851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.961053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.961060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.961284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.961293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.961595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.961602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.961901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.961908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.962218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.962225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.962393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.962400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.962439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.962445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.962487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.962494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.962789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.962796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.963107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.963114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.963290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.630 [2024-11-26 19:26:59.963297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.630 qpair failed and we were unable to recover it. 00:30:47.630 [2024-11-26 19:26:59.963678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.963685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.963986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.963993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.964335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.964342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.964511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.964518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.964840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.964847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.965222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.965229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.965387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.965395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.965609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.965615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.965953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.965960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.966252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.966259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.966419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.966425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.966708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.966716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.967096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.967103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.967393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.967399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.967789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.967796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.968087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.968094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.968412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.968419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.968646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.968654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.968988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.968995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.969337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.969343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.969665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.969672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.969892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.969899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.970200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.970207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.970367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.970374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.970678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.970684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.971001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.971008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.971185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.971192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.971474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.971480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.971880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.971887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.972172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.972178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.972495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.972504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.972835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.972842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.973022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.973029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.973316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.973322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.973545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.973552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.973887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.973895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.974083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.631 [2024-11-26 19:26:59.974090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.631 qpair failed and we were unable to recover it. 00:30:47.631 [2024-11-26 19:26:59.974432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.974439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.974752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.974758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.974928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.974935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.975218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.975224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.975533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.975540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.975742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.975749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.976061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.976068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.976394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.976400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.976575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.976583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.976785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.976792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.977112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.977119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.977416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.977423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.977575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.977581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.977869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.977876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.978189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.978196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.978489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.978496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.978764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.978771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.979113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.979120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.979539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.979546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.979859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.979869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.980209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.980216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.980539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.980545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.980714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.980722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.981026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.981033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.981395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.981402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.981698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.981704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.981874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.981880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.982176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.982183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.982457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.982464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.982815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.982821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.983013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.983020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.983310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.983317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.983614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.983620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.984024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.984033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.984350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.984357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.984681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.984688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.985017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.985024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.985210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.985216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.632 [2024-11-26 19:26:59.985383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.632 [2024-11-26 19:26:59.985390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.632 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.985671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.985677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.985868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.985874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.986174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.986182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.986346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.986354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.986509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.986516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.986790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.986797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.987101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.987107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.987398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.987404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.987620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.987628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.987826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.987832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.988141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.988148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.988469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.988476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.988662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.988669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.988844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.988850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.989143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.989150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.989329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.989336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.989657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.989664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.989980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.989987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.990294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.990300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.990496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.990512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.990700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.990707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.990977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.990984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.991324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.991331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.991655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.991663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.991989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.991996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.992303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.992309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.992589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.992597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.992924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.992931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.993270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.993276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.993591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.993598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.993809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.993815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.994048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.994056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.994380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.994387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.994604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.994611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.994784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.994793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.995105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.995112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.995334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.995341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.995643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.633 [2024-11-26 19:26:59.995649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.633 qpair failed and we were unable to recover it. 00:30:47.633 [2024-11-26 19:26:59.995826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.995832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.996022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.996029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.996196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.996202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.996450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.996456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.996761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.996768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.997061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.997068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.997362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.997369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.997525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.997532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.997848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.997855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.998186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.998193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.998469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.998476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.998811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.998818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.999104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.999111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.999277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.999285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.999584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.999591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:26:59.999867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:26:59.999874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.000208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.000215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.000532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.000539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.000899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.000906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.001270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.001277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.001499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.001507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.001577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.001583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.001964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.001971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.002601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.002616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.002789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.002797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.003043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.003050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.003432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.003439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.003761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.003768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.004001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.004010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.004051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.004060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.004367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.004373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.004535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.004541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.004776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.634 [2024-11-26 19:27:00.004783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.634 qpair failed and we were unable to recover it. 00:30:47.634 [2024-11-26 19:27:00.004982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.004990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.005312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.005319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.005626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.005634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.005834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.005844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.006002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.006009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.006308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.006315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.006505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.006513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.006801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.006808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.006895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.006902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.007099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.007107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.007294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.007342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.007670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.007678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.008002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.008009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.008208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.008215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.008257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.008265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.008589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.008596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.008909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.008916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.009104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.009113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.009503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.009511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.009819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.009827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.010063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.010071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.010410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.010417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.010591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.010599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.010698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.010706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.010927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.010935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.011304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.011311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.011473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.011481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.011852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.011859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.012027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.012035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.012362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.012369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.012663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.012671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.012857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.012867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.013197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.013204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.013399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.013415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.013768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.013775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.014096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.014104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.014435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.014443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.014607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.014614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.635 [2024-11-26 19:27:00.014814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.635 [2024-11-26 19:27:00.014823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.635 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.015079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.015086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.015256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.015264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.015627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.015634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.015826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.015833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.016022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.016030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.016109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.016116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.016287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.016295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.016408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.016416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.016740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.016748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.016827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.016835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.017008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.017015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.017182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.017190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.017301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.017308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.017561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.017568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.017906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.017913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.017954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.017961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.018028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.018035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.018268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.018275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.018581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.018588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.018950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.018957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.019325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.019332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.019507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.019514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.019800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.019807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.020070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.020078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.020347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.020355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.020598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.020606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.020783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.020791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.021106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.021113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.021309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.021316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.021386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.021393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.021603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.021610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.021914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.021925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.022108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.022116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.022347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.022355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.022536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.022544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.022624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.022630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.022799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.022807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.023062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.636 [2024-11-26 19:27:00.023070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.636 qpair failed and we were unable to recover it. 00:30:47.636 [2024-11-26 19:27:00.023255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.023263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.023559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.023566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.023873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.023881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.024057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.024065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.024372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.024379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.024559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.024567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.024855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.024867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.025230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.025237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.025531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.025538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.025837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.025844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.026073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.026081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.026372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.026379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.026653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.026660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.026985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.026992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.027221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.027228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.027460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.027468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.027508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.027515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.027848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.027856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.028036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.028044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.028346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.028353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.028543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.028550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.028795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.028803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.029111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.029119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.029411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.029417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.029624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.029631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.029959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.029967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.030269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.030277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.030443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.030450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.030727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.030734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.030885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.030893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.031137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.031144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.031461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.031468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.031646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.031654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.031940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.031951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.032186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.032192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.032367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.032374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.032535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.032542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.032868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.032875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.637 qpair failed and we were unable to recover it. 00:30:47.637 [2024-11-26 19:27:00.033229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.637 [2024-11-26 19:27:00.033237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.033596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.033603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.033775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.033783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.034105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.034112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.034278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.034285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.034521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.034528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.034835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.034841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.035158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.035166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.035351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.035358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.035582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.035589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.035766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.035774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.036073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.036080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.036301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.036308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.036493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.036500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.036575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.036583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.036662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.036670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.036982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.036990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.037307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.037314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.037649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.037655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.037994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.038001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.038322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.038329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.038607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.038615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.038659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.038666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.038861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.038873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.039157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.039164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.039458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.039465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.039767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.039775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.040092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.040100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.040427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.040434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.040726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.040733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.041010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.041017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.041212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.041219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.041530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.041537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.041598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.041604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.041696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.041703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.041854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.041867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.042168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.042176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.042487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.042494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.638 [2024-11-26 19:27:00.042827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.638 [2024-11-26 19:27:00.042834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.638 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.043129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.043138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.043314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.043322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.043671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.043679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.043887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.043894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.044201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.044207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.044401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.044409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.044668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.044675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.044717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.044725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.044958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.044966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.045299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.045306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.045349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.045356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.045570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.045578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.045771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.045779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.046026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.046034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.046295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.046302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.046462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.046469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.046762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.046769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.047000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.047008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.047170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.047176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.047257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.047264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.047348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.047355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.047565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.047573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.047917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.047925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.047993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.048000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.048090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.048097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.048356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.048364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.048719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.048727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.049051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.049059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.049280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.049289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.049474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.049481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.049788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.049796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.050166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.050174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.050461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.050469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.639 [2024-11-26 19:27:00.050530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.639 [2024-11-26 19:27:00.050537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.639 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.050825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.050833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.051013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.051022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.051239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.051248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.051445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.051453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.051797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.051804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.052113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.052121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.052294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.052301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.052536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.052543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.052851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.052858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.053173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.053180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.053373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.053380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.053619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.053626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.053665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.053672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.054036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.054044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.054188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.054195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.054375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.054382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.054710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.054718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.055077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.055085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.055299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.055307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.055623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.055630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.055978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.055985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.056167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.056174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.056600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.056607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.056909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.056916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.057176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.057183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.057439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.057446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.057750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.057756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.057945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.057953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.058316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.058323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.058468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.058475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.058679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.058685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.058865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.058873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.058968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.058975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.059158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.059166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.059457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.059464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.059772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.059779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.059944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.059952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.060337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.640 [2024-11-26 19:27:00.060344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.640 qpair failed and we were unable to recover it. 00:30:47.640 [2024-11-26 19:27:00.060653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.060660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.060942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.060949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.061130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.061137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.061324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.061331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.061369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.061377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.061556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.061563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.061927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.061935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.062135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.062143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.062456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.062463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.062632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.062640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.062915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.062922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.063262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.063269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.063452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.063459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.063681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.063688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.063989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.063996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.064328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.064335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.064640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.064647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.064944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.064951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.065330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.065337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.065756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.065763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.065928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.065935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.066292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.066299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.066359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.066366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.066533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.066540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.066822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.066829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.067142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.067150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.067359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.067367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.067555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.067562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.067859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.067871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.068196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.068204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.068541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.068548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.068835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.068843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.069208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.069215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.069545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.069552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.069855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.069866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.070149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.070156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.070329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.070335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.070533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.070539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.641 [2024-11-26 19:27:00.070884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.641 [2024-11-26 19:27:00.070893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.641 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.071183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.071190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.071519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.071525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.071733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.071740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.071926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.071933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.072121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.072128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.072282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.072291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.072536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.072543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.072721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.072728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.072892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.072899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.073115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.073122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.073570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.073577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.073743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.073750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.073957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.073965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.074262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.074269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.074432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.074439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.074722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.074729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.074911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.074919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.075264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.075271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.075430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.075438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.075721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.075728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.076050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.076057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.076228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.076235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.076515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.076522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.076701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.076709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.076999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.077006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.077065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.077071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.077349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.077356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.077679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.077686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.078003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.078010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.078213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.078219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.078458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.078465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.078595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.078601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.078787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.078794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.079090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.079097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.079433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.079440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.079772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.079780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.080113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.080121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.080440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.642 [2024-11-26 19:27:00.080448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.642 qpair failed and we were unable to recover it. 00:30:47.642 [2024-11-26 19:27:00.080809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.080816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.081033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.081041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.081420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.081427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.081603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.081610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.081912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.081920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.082141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.082148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.082384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.082392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.082697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.082706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.082879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.082886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.083169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.083176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.083350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.083363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.083577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.083585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.083868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.083875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.084168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.084175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.084464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.084471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.084800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.084807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.084976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.084984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.085159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.085166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.085468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.085475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.085513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.085520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.085875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.085882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.086163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.086169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.086352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.086359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.086528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.086535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.086717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.086725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.087083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.087091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.087283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.087291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.087591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.087598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.087917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.087924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.087959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.087965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.088154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.088161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.088502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.088509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.088691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.088699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.643 [2024-11-26 19:27:00.088919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.643 [2024-11-26 19:27:00.088926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.643 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.089135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.089143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.089499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.089506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.089803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.089810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.089992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.090000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.090289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.090296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.090366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.090372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.090558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.090565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.090868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.090875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.091226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.091233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.091403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.091411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.091724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.091731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.092083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.092090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.092440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.092447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.092763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.092773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.093105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.093113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.093290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.093297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.093462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.093470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.093666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.093674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.093986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.093993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.094331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.094339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.094671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.094679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.094853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.094861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.095260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.095268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.095388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.095396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.095579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.095586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.095783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.095790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.096092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.096099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.096412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.096420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.096720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.096729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.097034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.097042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.097216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.097225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.097527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.097535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.097840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.097848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.098029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.098036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.098392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.098399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.098734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.098741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.098899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.098907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.099084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.644 [2024-11-26 19:27:00.099091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.644 qpair failed and we were unable to recover it. 00:30:47.644 [2024-11-26 19:27:00.099334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.099341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.099694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.099702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.099889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.099897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.100179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.100185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.100365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.100373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.100563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.100570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.100762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.100776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.101059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.101066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.101374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.101381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.101676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.101683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.102056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.102063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.102170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.102177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.102520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.102527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.102767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.102774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.102992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.103001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.103299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.103307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.103489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.103496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.103715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.103721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.103942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.103950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.104163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.104169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.104381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.104388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.104756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.104762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.104984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.104991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.105308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.105315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.105494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.105502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.105695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.105702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.105898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.105905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.106105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.106112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.106402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.106409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.106814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.106821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.107116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.107123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.107419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.107426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.107598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.107604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.107926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.107934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.108122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.108129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.108449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.108456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.108744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.108751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.109043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.109051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.645 qpair failed and we were unable to recover it. 00:30:47.645 [2024-11-26 19:27:00.109396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.645 [2024-11-26 19:27:00.109404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.109716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.109723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.109894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.109901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.110110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.110117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.110492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.110499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.110819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.110825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.111168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.111175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.111383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.111390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.111645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.111653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.111812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.111820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.112150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.112158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.112514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.112521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.112747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.112754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.113118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.113125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.113428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.113435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.113473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.113479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.113659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.113666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.114001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.114010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.114239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.114246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.114576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.114582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.114898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.114905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.115219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.115226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.115436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.115443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.115767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.115774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.116104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.116111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.116275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.116283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.116693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.116700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.116879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.116887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.117050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.117057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.117223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.117230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.117457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.117464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.117807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.117814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.118023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.118031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.118192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.118199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.118360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.118367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.118592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.118600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.118648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.118656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.118961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.118968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.646 [2024-11-26 19:27:00.119280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.646 [2024-11-26 19:27:00.119287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.646 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.119606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.119613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.119809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.119816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.119976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.119983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.120341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.120348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.120641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.120648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.120847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.120866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.121158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.121165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.121485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.121492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.121815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.121822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.122027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.122034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.122341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.122348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.122545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.122553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.122595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.122603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.122956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.122963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.123348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.123354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.123549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.123556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.123758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.123764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.123935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.123942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.124295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.124303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.124650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.124657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.124955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.124962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.125146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.125153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.125534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.125541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.125881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.125888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.126253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.126260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.126472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.126479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.126658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.126665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.126881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.126888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.127052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.127059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.127230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.127237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.127539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.127545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.127833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.127840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.128156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.128163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.128201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.128207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.128367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.128374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.128744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.128751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.128933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.128940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.129216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.647 [2024-11-26 19:27:00.129222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.647 qpair failed and we were unable to recover it. 00:30:47.647 [2024-11-26 19:27:00.129517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.129523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.129817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.129823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.130007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.130015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.130408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.130415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.130728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.130736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.130943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.130950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.131260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.131267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.131590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.131597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.131936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.131944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.132135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.132142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.132298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.132304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.132611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.132617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.132939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.132947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.133137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.133144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.133183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.133190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.133497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.133503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.133685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.133692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.134055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.134062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.134358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.134365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.134681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.134688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.135031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.135040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.135254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.135260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.135592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.135598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.135885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.135892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.136095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.136103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.136426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.136433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.136822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.136829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.137006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.137014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.137321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.137327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.137712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.137719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.138035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.138042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.138263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.138270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.138582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.138589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.138904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.648 [2024-11-26 19:27:00.138911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.648 qpair failed and we were unable to recover it. 00:30:47.648 [2024-11-26 19:27:00.139089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.139096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.139375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.139381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.139700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.139707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.140041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.140048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.140388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.140394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.140689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.140696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.141012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.141020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.141431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.141439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.141612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.141620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.141794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.141802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.142173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.142180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.142377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.142385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.142572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.142579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.142901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.142909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.143120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.143132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.143472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.143479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.143771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.143778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.144171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.144179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.144294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.144300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.144496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.144502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.144662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.144670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.144971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.144978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.145274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.145281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.145320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.145328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.145631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.145638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.146014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.146021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.146323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.146330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.146506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.146514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.146920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.146927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.147094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.147102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.147289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.147296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.147590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.147597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.147961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.147969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.148284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.148291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.148457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.148463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.148704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.148711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.148966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.148973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.149293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.649 [2024-11-26 19:27:00.149300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.649 qpair failed and we were unable to recover it. 00:30:47.649 [2024-11-26 19:27:00.149580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.149588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.149747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.149754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.150050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.150057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.150355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.150362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.150656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.150663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.150981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.150988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.151166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.151173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.151340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.151346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.151549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.151555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.151752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.151758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.152015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.152023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.152424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.152432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.152740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.152748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.153064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.153071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.153252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.153260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.153578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.153586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.153731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.153737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.154053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.154060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.154275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.154281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.154565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.154571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.154918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.154925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.155249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.155256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.155553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.155560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.155739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.155747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.156060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.156068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.156377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.156384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.156685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.156692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.156998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.157005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.157321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.157329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.157539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.157547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.157850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.157857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.158034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.158041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.158344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.158351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.158678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.158684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.159000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.159007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.159226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.159233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.159622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.159629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.159699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.159705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.650 [2024-11-26 19:27:00.159881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.650 [2024-11-26 19:27:00.159887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.650 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.160186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.160192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.160555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.160562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.160870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.160878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.161253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.161260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.161433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.161440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.161763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.161769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.162070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.162077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.162452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.162458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.162641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.162648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.162830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.162837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.163086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.163094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.163413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.163419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.163644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.163650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.163841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.163848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.164151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.164158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.164514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.164520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.164592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.164600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.164839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.164846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.165044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.165051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.165244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.165251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.165533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.165540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.165941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.165948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.166343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.166350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.166512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.166520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.166601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.166608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.166799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.166807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.167147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.167154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.167193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.167200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.167350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.167357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.167719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.167725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.168151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.168158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.168362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.168369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.168513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.168519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.168855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.168864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.169104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.169111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.169429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.169436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.169647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.169654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.169880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.651 [2024-11-26 19:27:00.169887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.651 qpair failed and we were unable to recover it. 00:30:47.651 [2024-11-26 19:27:00.170258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.170264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.170559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.170565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.170882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.170890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.171194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.171200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.171602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.171609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.171925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.171932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.172260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.172267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.172396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.172402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.172711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.172717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.172891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.172898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.173082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.173089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.173306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.173312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.173628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.173634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.173802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.173808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.174131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.174139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.174305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.174313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.174517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.174525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.174813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.174820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.175034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.175043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.175351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.175358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.175524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.175531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.175686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.175692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.175965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.175972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.176302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.176308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.176531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.176538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.176741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.176748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.177071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.177078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.177408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.177416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.177753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.177760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.178077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.178085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.178418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.178425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.178807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.178813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.179128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.179136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.179455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.652 [2024-11-26 19:27:00.179462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.652 qpair failed and we were unable to recover it. 00:30:47.652 [2024-11-26 19:27:00.179777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.179784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.179970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.179977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.180190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.180196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.180520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.180526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.180870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.180878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.181203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.181209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.181612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.181619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.181936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.181943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.182121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.182128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.182359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.182365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.182800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.182807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.183116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.183123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.183431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.183437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.183596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.183603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.183889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.183896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.184218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.184226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.184422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.184429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.184572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.184579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.184904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.184912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.185236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.185243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.185538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.185545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.185844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.185850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.186166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.186173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.186343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.186349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.186656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.186664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.186934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.186941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.187169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.187175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.187386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.187394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.187578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.187586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.187952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.187959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.188141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.188148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.188437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.188443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.188637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.188644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.188961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.188968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.189323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.189330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.189646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.189652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.189965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.189972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.190415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.190421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.653 qpair failed and we were unable to recover it. 00:30:47.653 [2024-11-26 19:27:00.190742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.653 [2024-11-26 19:27:00.190749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.191063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.191070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.191393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.191401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.191574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.191582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.191905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.191913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.192188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.192194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.192587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.192593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.192932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.192939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.193144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.193150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.193326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.193332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.193530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.193537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.193719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.193727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.193942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.193948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.194145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.194153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.194434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.194441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.194798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.194806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.194981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.194989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.195193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.195200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.195536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.195543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.195829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.195835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.196167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.196174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.196214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.196221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.196590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.196596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.196930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.196937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.197317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.197325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.197619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.197626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.197936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.197947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.198128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.198135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.198513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.198520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.198805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.198812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.199147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.199155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.199468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.199475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.199773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.199780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.200083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.200090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.200397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.200404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.200484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.200491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.200785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.200791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.200968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.200976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.654 [2024-11-26 19:27:00.201173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.654 [2024-11-26 19:27:00.201179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.654 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.201503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.201510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.201802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.201809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.202183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.202190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.202356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.202362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.202404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.202411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.202834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.202841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.203009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.203017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.203300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.203306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.203491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.203499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.203814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.203820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.203976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.203983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.204029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.204035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.204228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.204235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.204439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.204446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.204610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.204617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.204924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.204931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.205112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.205118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.205405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.205412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.205719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.205725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.205898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.205911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.206071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.206078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.206377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.206385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.206692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.206699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.206916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.206924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.207259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.207266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.207557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.207564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.207735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.207742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.208063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.208072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.208380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.208387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.208729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.208735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.209057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.209064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.209248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.209256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.209573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.209579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.209890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.209897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.210084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.210092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.210252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.210259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.210613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.210620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.210938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.655 [2024-11-26 19:27:00.210945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.655 qpair failed and we were unable to recover it. 00:30:47.655 [2024-11-26 19:27:00.211291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.211299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.211595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.211601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.211774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.211782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.212130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.212137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.212430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.212437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.212643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.212650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.212972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.212979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.213431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.213438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.213786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.213794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.214090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.214097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.214263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.214271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.214553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.214560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.214874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.214881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.215215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.215222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.215536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.215543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.215744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.215751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.215925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.215932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.216098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.216105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.216410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.216417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.216706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.216712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.216923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.216932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.217254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.217262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.217436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.217444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.217827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.217835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.218216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.218224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.218378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.218385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.218681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.218689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.218838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.218845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.219129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.219137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.219524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.219533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.219820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.219826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.220011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.220019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.220372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.220378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.220670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.220677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.220886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.220893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.221175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.221181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.221400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.221406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.656 [2024-11-26 19:27:00.221615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.656 [2024-11-26 19:27:00.221622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.656 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.221928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.221936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.222265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.222274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.222446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.222453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.222757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.222764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.222933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.222941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.223240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.223247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.223558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.223565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.223838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.223846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.224145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.224152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.224313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.224320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.224651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.224658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.225004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.225011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.225206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.225213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.225583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.225589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.225751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.225766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.226069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.226077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.945 [2024-11-26 19:27:00.226290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.945 [2024-11-26 19:27:00.226297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.945 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.226590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.226597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.226753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.226760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.227014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.227022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.227314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.227321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.227633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.227639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.227832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.227839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.228124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.228131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.228439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.228446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.228567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.228573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.228904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.228911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.229225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.229233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.229413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.229420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.229577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.229585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.229779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.229786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.230091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.230100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.230393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.230399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.230700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.230707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.231035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.231042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.231365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.231371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.231696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.231703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.232045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.232053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.232393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.232400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.232592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.232598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.232786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.232793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.233144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.233151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.233326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.233333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.233509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.233516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.233791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.233798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.234160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.234167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.234491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.234498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.234798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.234804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.234928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.234935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.235168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.235175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.235338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.235345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.235644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.235651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.235949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.235956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.236125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.946 [2024-11-26 19:27:00.236133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.946 qpair failed and we were unable to recover it. 00:30:47.946 [2024-11-26 19:27:00.236421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.236428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.236601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.236607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.236924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.236931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.237272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.237280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.237628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.237635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.237702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.237708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.238020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.238027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.238431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.238438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.238603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.238610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.238822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.238830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.239168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.239176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.239496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.239504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.239693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.239701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.239770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.239778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.240080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.240087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.240407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.240415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.240577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.240584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.241034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.241043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.241400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.241407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.241567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.241574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.241807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.241815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.241993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.242000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.242306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.242313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.242496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.242503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.242787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.242794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.243159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.243166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.243489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.243496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.243839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.243846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.244138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.244145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.244439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.244447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.244797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.244805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.244988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.244995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.245284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.245290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.245583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.245589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.245912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.245919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.246223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.246230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.246391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.246399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.246708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.947 [2024-11-26 19:27:00.246716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.947 qpair failed and we were unable to recover it. 00:30:47.947 [2024-11-26 19:27:00.246883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.246891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.247193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.247200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.247523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.247530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.247855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.247864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.248045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.248051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.248341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.248348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.248658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.248664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.248824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.248831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.249144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.249151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.249524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.249531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.249949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.249956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.250275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.250283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.250607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.250614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.250921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.250928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.251114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.251121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.251532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.251538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.251841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.251848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.252084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.252091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.252410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.252417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.252742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.252752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.253207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.253214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.253533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.253540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.253874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.253882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.254182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.254189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.254508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.254514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.254838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.254845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.255033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.255040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.255281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.255287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.255492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.255500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.255858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.255868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.256063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.256070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.256402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.256409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.256570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.256576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.256873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.256880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.257262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.257269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.257595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.257602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.257942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.257949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.258251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.948 [2024-11-26 19:27:00.258257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.948 qpair failed and we were unable to recover it. 00:30:47.948 [2024-11-26 19:27:00.258476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.258483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.258811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.258818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.259002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.259010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.259343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.259350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.259504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.259511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.259843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.259849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.260202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.260209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.260554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.260562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.260910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.260917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.261212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.261220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.261534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.261540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.261923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.261930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.262262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.262268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.262447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.262454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.262737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.262744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.263034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.263041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.263234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.263241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.263551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.263557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.263880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.263887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.264186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.264192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.264526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.264534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.264812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.264820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.265257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.265264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.265554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.265561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.265871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.265878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.266199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.266205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.266512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.266518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.266843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.266849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.267149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.267156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.267466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.267473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.267804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.267811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.268120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.268128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.268294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.268301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.268335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.268341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.268664] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.268671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.268851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.268859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.269151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.269158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.269350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.269357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.949 [2024-11-26 19:27:00.269716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.949 [2024-11-26 19:27:00.269723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.949 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.270022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.270029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.270362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.270368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.270557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.270564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.270780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.270786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.271081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.271088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.271415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.271422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.271786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.271793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.271991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.271999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.272182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.272189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.272515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.272523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.272840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.272847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.273031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.273039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.273324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.273331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.273752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.273759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.274056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.274063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.274353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.274361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.274670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.274678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.274851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.274859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.275164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.275171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.275526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.275533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.275689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.275697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.275878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.275886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.276065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.276074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.276379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.276386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.276551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.276560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.276727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.276734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.277030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.277038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.277370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.277377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.277706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.277714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.277948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.277956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.278320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.278327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.278641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.278648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.950 [2024-11-26 19:27:00.278825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.950 [2024-11-26 19:27:00.278832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.950 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.279208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.279216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.279520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.279527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.279685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.279692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.280044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.280052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.280357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.280364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.280655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.280661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.280969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.280976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.281278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.281285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.281476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.281483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.281663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.281671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.282022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.282030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.282309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.282316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.282632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.282639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.282957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.282965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.283151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.283158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.283444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.283451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.283617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.283624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.283910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.283916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.284099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.284106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.284403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.284410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.284607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.284623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.284974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.284981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.285285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.285291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.285636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.285643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.285819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.285827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.286115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.286122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.286415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.286422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.286577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.286583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.286944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.286951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.287241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.287248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.287563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.287570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.287900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.287907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.288124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.288131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.288503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.288510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.288678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.288685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.288871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.288877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.289050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.289056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.289342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.951 [2024-11-26 19:27:00.289349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.951 qpair failed and we were unable to recover it. 00:30:47.951 [2024-11-26 19:27:00.289659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.289666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.289970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.289977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.290199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.290206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.290375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.290382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.290658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.290666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.290855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.290866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.291071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.291078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.291310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.291317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.291496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.291502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.291777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.291784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.292161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.292168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.292476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.292483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.292658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.292665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.292826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.292834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.293051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.293058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.293254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.293261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.293561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.293568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.293618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.293626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.293976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.293984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.294298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.294304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.294642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.294649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.294846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.294852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.295171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.295178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.295457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.295464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.295615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.295622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.295802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.295809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.296035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.296043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.296217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.296223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.296573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.296580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.296754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.296761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.296989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.296996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.297061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.297068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.297370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.297376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.297592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.297607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.297923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.297930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.298258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.298265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.298534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.298541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.298853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.298860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.952 qpair failed and we were unable to recover it. 00:30:47.952 [2024-11-26 19:27:00.299064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.952 [2024-11-26 19:27:00.299072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.299412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.299419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.299601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.299609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.299958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.299965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.300222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.300229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.300408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.300415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.300584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.300592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.300970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.300977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.301299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.301306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.301638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.301645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.301858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.301869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.302071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.302078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.302275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.302288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.302644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.302652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.302827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.302835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.303072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.303079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.303256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.303264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.303647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.303653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.303981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.303987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.304163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.304170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.304449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.304458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.304655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.304662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.304856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.304869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.305255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.305262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.305580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.305588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.305760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.305767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.306079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.306087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.306250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.306258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.306616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.306623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.306667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.306674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.306991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.306998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.307356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.307363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.307529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.307536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.307699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.307706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.307867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.307874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.307944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.307951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.308138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.308145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.308478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.308485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.308783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.953 [2024-11-26 19:27:00.308790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.953 qpair failed and we were unable to recover it. 00:30:47.953 [2024-11-26 19:27:00.309091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.309098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.309321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.309327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.309503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.309510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.309696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.309702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.310019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.310026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.310360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.310367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.310699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.310707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.310900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.310908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.311144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.311151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.311488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.311495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.311676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.311683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.312014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.312021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.312343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.312349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.312629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.312635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.312700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.312706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.312986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.312993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.313320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.313327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.313645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.313651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.313953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.313960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.314295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.314301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.314484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.314493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.314778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.314787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.315140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.315147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.315514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.315521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.315888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.315895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.316211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.316218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.316537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.316544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.316722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.316730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.317142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.317149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.317323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.317332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.317626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.317634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.317951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.317958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.318291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.318298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.954 qpair failed and we were unable to recover it. 00:30:47.954 [2024-11-26 19:27:00.318640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.954 [2024-11-26 19:27:00.318647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.318874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.318881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.319247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.319254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.319569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.319575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.319764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.319772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.319992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.319999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.320191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.320199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.320544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.320551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.320790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.320797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.320994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.321001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.321076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.321083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.321396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.321403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.321565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.321572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.321652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.321660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.322021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.322028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.322190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.322198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.322483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.322489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.322675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.322682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.322843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.322850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.323218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.323225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.323555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.323563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.323932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.323940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.324308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.324315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.324633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.324640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.325027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.325034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.325224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.325236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.325574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.325581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.325912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.325919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.326129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.326139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.326483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.326490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.326662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.326669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.327036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.327043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.327355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.327362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.327572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.327580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.327880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.327887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.328164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.328171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.328503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.328511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.328810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.328817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.955 [2024-11-26 19:27:00.329148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.955 [2024-11-26 19:27:00.329156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.955 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.329479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.329486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.329667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.329673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.329964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.329972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.330147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.330155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.330317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.330324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.330588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.330596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.330766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.330774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.330967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.330974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.331281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.331288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.331585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.331592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.331762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.331769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.331922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.331931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.332118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.332125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.332461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.332467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.332867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.332874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.333209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.333216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.333407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.333415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.333745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.333752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.334067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.334074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.334383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.334389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.334429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.334436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.334793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.334801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.335196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.335203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.335461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.335468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.335799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.335805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.336133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.336141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.336324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.336332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.336632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.336639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.336819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.336827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.337140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.337149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.337345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.337351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.337501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.337508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.337832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.337839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.337969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.337976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.338351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.338357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.338672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.338679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.338980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.338987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.339278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.956 [2024-11-26 19:27:00.339285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.956 qpair failed and we were unable to recover it. 00:30:47.956 [2024-11-26 19:27:00.339576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.339583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.339739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.339746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.339901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.339909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.340216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.340223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.340568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.340575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.340871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.340878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.341037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.341044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.341332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.341339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.341637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.341643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.341818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.341825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.342136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.342143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.342487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.342493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.342796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.342802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.343133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.343140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.343301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.343308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.343472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.343478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.343652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.343659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.343950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.343957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.344135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.344143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.344455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.344462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.344665] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.344673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.344981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.344989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.345305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.345312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.345486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.345493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.345692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.345701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.346049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.346057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.346372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.346378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.346674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.346681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.346856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.346872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.347157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.347165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.347488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.347496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.347795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.347805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.347991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.347998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.348284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.348291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.348476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.348482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.348786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.348793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.349000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.349007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.349185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.957 [2024-11-26 19:27:00.349191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.957 qpair failed and we were unable to recover it. 00:30:47.957 [2024-11-26 19:27:00.349473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.349480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.349542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.349548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.349932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.349939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.350248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.350255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.350419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.350426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.350721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.350728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.351045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.351052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.351382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.351389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.351698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.351706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.351889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.351897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.352238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.352245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.352537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.352543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.352848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.352854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.353160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.353167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.353467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.353474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.353798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.353805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.353976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.353983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.354348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.354355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.354666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.354673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.354977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.354985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.355318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.355326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.355638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.355645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.355967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.355974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.356150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.356157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.356525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.356531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.356839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.356846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.357009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.357017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.357183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.357190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.357503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.357509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.357691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.357698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.358038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.358045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.358430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.358437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.358743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.358750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.358923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.358939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.359303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.359310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.359630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.359637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.360022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.360029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.360328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.360335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.958 [2024-11-26 19:27:00.360667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.958 [2024-11-26 19:27:00.360674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.958 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.360989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.360996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.361284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.361292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.361507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.361515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.361674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.361682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.361920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.361927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.362245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.362252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.362578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.362585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.362757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.362766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.362927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.362934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.363232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.363240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.363405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.363412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.363509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.363517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.363654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.363662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.363935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.363944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.364260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.364268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.364589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.364597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.364930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.364937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.365180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.365187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.365586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.365593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.365770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.365779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.365965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.365971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.366304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.366311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.366487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.366495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.366792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.366799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.367164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.367171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.367488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.367495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.367831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.367838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.368166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.368174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.368209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.368216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.368540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.368547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.368845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.368851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.369155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.369163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.369561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.369569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.369868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.369877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.959 qpair failed and we were unable to recover it. 00:30:47.959 [2024-11-26 19:27:00.370079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.959 [2024-11-26 19:27:00.370088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.370438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.370444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.370771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.370779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.370950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.370958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.371339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.371346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.371638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.371645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.371835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.371842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.372156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.372163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.372477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.372484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.372875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.372882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.373058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.373066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.373356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.373363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.373672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.373678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.373974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.373982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.374308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.374315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.374528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.374536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.374885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.374892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.375254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.375260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.375558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.375565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.375726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.375733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.376110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.376118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.376485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.376492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.376808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.376815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.377110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.377117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.377324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.377331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.377519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.377526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.377907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.377915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.378221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.378228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.378402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.378409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.378648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.378656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.378851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.378858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.379288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.379296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.379503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.379510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.379801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.379809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.379976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.379983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.380152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.380160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.380370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.380376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.380559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.380567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.960 [2024-11-26 19:27:00.380771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.960 [2024-11-26 19:27:00.380778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.960 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.381072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.381079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.381412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.381422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.381751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.381758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.382155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.382163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.382328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.382335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.382490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.382497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.382804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.382811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.382991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.382999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.383224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.383231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.383415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.383423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.383761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.383769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.384096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.384104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.384419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.384426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.384596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.384604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.384784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.384791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.385206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.385213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.385520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.385527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.385679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.385686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.385970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.385977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.386141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.386148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.386521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.386528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.386685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.386692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.386972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.386979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.387165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.387173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.387469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.387476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.387643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.387651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.387960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.387967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.388303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.388310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.388639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.388646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.388969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.388977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.389145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.389152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.389437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.389444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.389773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.389781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.390081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.390088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.390377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.390384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.390421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.390429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.390600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.390607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.961 [2024-11-26 19:27:00.390803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.961 [2024-11-26 19:27:00.390811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.961 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.390984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.390991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.391269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.391276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.391495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.391502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.391820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.391830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.392153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.392161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.392460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.392467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.392642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.392649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.392841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.392847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.393039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.393046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.393351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.393357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.393683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.393690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.394002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.394010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.394178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.394186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.394496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.394504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.394820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.394827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.395189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.395196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.395512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.395518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.395842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.395850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.396171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.396179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.396486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.396494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.396656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.396664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.396819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.396827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.397100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.397108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.397147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.397153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.397312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.397319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.397508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.397515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.397808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.397815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.398002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.398010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.398205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.398212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.398524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.398531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.398842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.398849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.399164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.399172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.399510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.399517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.399817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.399824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.399989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.399997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.400172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.400179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.400470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.400476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.400644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.962 [2024-11-26 19:27:00.400651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.962 qpair failed and we were unable to recover it. 00:30:47.962 [2024-11-26 19:27:00.400884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.400892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.401054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.401061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.401355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.401362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.401690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.401697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.401996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.402004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.402184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.402193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.402574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.402581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.402742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.402750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.402913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.402920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.403225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.403232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.403639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.403647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.403684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.403691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.403765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.403772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.404085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.404093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.404277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.404285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.404550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.404557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.404700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.404707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.404868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.404875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.405092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.405099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.405439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.405447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.405760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.405769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.406150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.406157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.406323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.406332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.406502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.406509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.406692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.406700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.406959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.406967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.407261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.407268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.407464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.407472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.407549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.407557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.407823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.407831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.408204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.408212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.408504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.408511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.408822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.408829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.409077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.409084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.963 qpair failed and we were unable to recover it. 00:30:47.963 [2024-11-26 19:27:00.409452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.963 [2024-11-26 19:27:00.409459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.409756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.409763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.409917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.409925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.410197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.410205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.410512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.410520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.410712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.410720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.410913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.410921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.411208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.411214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.411527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.411534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.411703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.411711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.411876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.411883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.412179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.412188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.412523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.412530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.412888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.412896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.413246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.413253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.413563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.413571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.413755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.413763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.414146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.414154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.414359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.414367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.414710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.414717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.415031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.415038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.415274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.415281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.415618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.415626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.415972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.415980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.416159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.416167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.416413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.416420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.416703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.416710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.417048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.417055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.417241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.417249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.417441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.417448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.417532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.417538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.417725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.417733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.417920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.417928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.418334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.418341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.418639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.418646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.419040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.419048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.419223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.419230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.419395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.964 [2024-11-26 19:27:00.419403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.964 qpair failed and we were unable to recover it. 00:30:47.964 [2024-11-26 19:27:00.419634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.419642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.420023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.420031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.420201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.420208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.420499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.420507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.420684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.420692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.421053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.421061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.421229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.421237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.421524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.421532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.421846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.421854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.422171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.422178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.422469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.422476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.422752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.422759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.423091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.423099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.423436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.423444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.423627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.423635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.423986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.423994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.424300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.424307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.424622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.424630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.424799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.424807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.425112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.425120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.425423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.425430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.425723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.425730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.426025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.426033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.426413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.426420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.426697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.426703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.426746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.426753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.427080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.427087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.427436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.427444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.427757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.427765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.428097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.428105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.428403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.428410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.428726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.428733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.428922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.428930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.429152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.429159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.429391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.429398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.429666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.429673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.429990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.429997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.430300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.430308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.965 qpair failed and we were unable to recover it. 00:30:47.965 [2024-11-26 19:27:00.430634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.965 [2024-11-26 19:27:00.430642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.430841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.430849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.430891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.430900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.431169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.431176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.431393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.431401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.431724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.431731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.432133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.432141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.432306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.432314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.432604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.432612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.432785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.432793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.433096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.433103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.433395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.433402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.433561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.433569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.433784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.433791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.433982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.433991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.434236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.434243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.434547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.434555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.434719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.434727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.435000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.435008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.435183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.435192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.435356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.435364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.435680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.435687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.435982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.435990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.436282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.436289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.436541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.436548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.436868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.436875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.437177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.437184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.437345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.437352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.437634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.437642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.437931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.437939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.438117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.438124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.438422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.438429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.438720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.438728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.439044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.439053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.439372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.439380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.439588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.439596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.439778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.439794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.440147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.440154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.966 [2024-11-26 19:27:00.440322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.966 [2024-11-26 19:27:00.440330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.966 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.440494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.440501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.440681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.440689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.441010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.441017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.441312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.441321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.441540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.441548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.441871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.441879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.442206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.442213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.442550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.442557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.442877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.442885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.443193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.443200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.443509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.443517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.443693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.443701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.443882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.443890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.444083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.444091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.444396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.444403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.444574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.444581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.444798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.444805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.445180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.445187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.445504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.445512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.445888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.445896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.446215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.446221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.446437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.446444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.446723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.446730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.446888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.446896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.447330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.447337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.447676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.447683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.447987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.447994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.448398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.448406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.448685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.448693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.448875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.448883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.449253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.449261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.449441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.449449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.449739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.449746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.449935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.449943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.450081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.450088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.450263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.450271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.450559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.450566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.450896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.967 [2024-11-26 19:27:00.450904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.967 qpair failed and we were unable to recover it. 00:30:47.967 [2024-11-26 19:27:00.451086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.451092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.451402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.451409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.451798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.451806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.452099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.452107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.452423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.452430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.452735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.452744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.452913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.452922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.453201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.453207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.453502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.453510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.453670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.453678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.453963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.453970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.454192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.454199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.454503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.454511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.454881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.454889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.455174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.455181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.455479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.455486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.455783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.455790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.456163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.456171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.456212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.456220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.456496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.456503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.456821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.456829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.457146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.457155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.457359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.457366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.457573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.457580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.457870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.457878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.458212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.458220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.458555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.458562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.458777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.458793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.459121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.459128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.459300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.459308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.459661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.459670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.459807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.459814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.460002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.460009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.460312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.460319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.968 [2024-11-26 19:27:00.460484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.968 [2024-11-26 19:27:00.460491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.968 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.460775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.460782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.461096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.461103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.461273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.461280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.461637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.461644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.461947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.461954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.462139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.462147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.462473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.462480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.462661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.462668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.462868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.462876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.462964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.462971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.463140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.463149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.463486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.463493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.463646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.463653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.463843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.463849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.463947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.463954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.464114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.464121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.464435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.464442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.464612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.464618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.464821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.464828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.465131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.465139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.465324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.465331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.465639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.465647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.465808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.465816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.466108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.466115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.466316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.466323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.466517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.466524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.466815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.466823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.467030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.467038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.467316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.467324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.467530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.467537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.467812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.467819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.468159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.468167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.468465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.468471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.468650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.468658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.468955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.468962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.469322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.469328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.469693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.469700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.470003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.969 [2024-11-26 19:27:00.470012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.969 qpair failed and we were unable to recover it. 00:30:47.969 [2024-11-26 19:27:00.470373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.470380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.470684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.470692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.471015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.471022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.471330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.471337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.471743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.471750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.471907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.471914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.472161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.472169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.472487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.472494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.472697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.472705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.472910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.472917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.473287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.473294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.473625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.473632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.473968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.473978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.474196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.474203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.474421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.474427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.474585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.474592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.474878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.474885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.475252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.475258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.475582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.475589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.475765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.475772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.476052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.476059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.476371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.476378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.476724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.476731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.477083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.477091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.477272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.477279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.477587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.477594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.477930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.477938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.477981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.477988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.478340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.478348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.478650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.478657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.478971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.478978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.479152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.479160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.479477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.479484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.479701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.479708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.479923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.479931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.480171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.480178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.480369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.480376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.480601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.480607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.970 qpair failed and we were unable to recover it. 00:30:47.970 [2024-11-26 19:27:00.480969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.970 [2024-11-26 19:27:00.480976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.481304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.481311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.481489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.481496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.481833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.481841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.482017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.482024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.482337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.482345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.482505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.482512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.482808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.482816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.483253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.483261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.483458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.483464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.483690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.483696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.483923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.483930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.484109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.484116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.484190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.484196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.484587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.484596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.484901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.484909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.485217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.485224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.485553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.485560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.485776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.485784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:47.971 [2024-11-26 19:27:00.486009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.486017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@868 -- # return 0 00:30:47.971 [2024-11-26 19:27:00.486284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.486292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:30:47.971 [2024-11-26 19:27:00.486664] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.486671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:30:47.971 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:47.971 [2024-11-26 19:27:00.487045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.487052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.487350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.487357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.487555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.487562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.487884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.487892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.488199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.488206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.488364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.488371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.488518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.488524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.488812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.488820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.489176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.489183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.489337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.489344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.489499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.489506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.489825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.489832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.490050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.490058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.490414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.490422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.971 [2024-11-26 19:27:00.490772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.971 [2024-11-26 19:27:00.490781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.971 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.490950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.490957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.491152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.491159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.491466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.491475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.491807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.491814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.492023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.492031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.492330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.492338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.492372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.492380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.492667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.492674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.492979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.492986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.493282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.493289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.493462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.493469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.493779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.493786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.493966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.493974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.494272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.494282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.494442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.494449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.494720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.494728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.495040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.495047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.495263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.495271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.495654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.495662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.495955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.495963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.496393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.496400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.496692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.496700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.496809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.496816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.497120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.497127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.497312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.497320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.497491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.497499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.497817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.497824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.498133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.498140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.498427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.498434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.498757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.498763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.498955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.498964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.499310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.499317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.499598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.499606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.499945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.972 [2024-11-26 19:27:00.499952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.972 qpair failed and we were unable to recover it. 00:30:47.972 [2024-11-26 19:27:00.500146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.500154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.500490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.500498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.500772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.500779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.500961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.500968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.501340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.501347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.501525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.501532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.501904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.501911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.502322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.502329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.502486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.502493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.502742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.502750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.503039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.503047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.503359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.503367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.503634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.503641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.503960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.503967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.504185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.504193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.504388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.504395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.504685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.504691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.504756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.504763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.504945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.504953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.505271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.505278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.505574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.505580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.505756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.505766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.506013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.506020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.506351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.506358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.506657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.506664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.507002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.507011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.507318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.507326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.507518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.507526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.507831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.507838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.508103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.508111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.508292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.508299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.508450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.508457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.508740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.508747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.509136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.509144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.509328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.509336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.509647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.509654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.509822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.509830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.510181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.973 [2024-11-26 19:27:00.510188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.973 qpair failed and we were unable to recover it. 00:30:47.973 [2024-11-26 19:27:00.510552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.510558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.510760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.510767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.511210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.511217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.511383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.511391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.511683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.511691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.512016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.512024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.512207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.512214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.512252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.512259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.512585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.512592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.512788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.512795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.513133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.513141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.513295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.513302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.513715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.513722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.514045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.514053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.514232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.514240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.514522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.514530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.514698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.514705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.514964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.514971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.515247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.515255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.515580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.515587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.515920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.515928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.516109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.516116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.516485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.516493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.516809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.516819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.517112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.517120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.517461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.517468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.517753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.517762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.518000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.518008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.518270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.518277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.518444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.518451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.518609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.518617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.518917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.518925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.519237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.519244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.519400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.519407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.519785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.519793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.519867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.519873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.520207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.520214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.520379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.974 [2024-11-26 19:27:00.520387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.974 qpair failed and we were unable to recover it. 00:30:47.974 [2024-11-26 19:27:00.520555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.520562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.520947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.520955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.521267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.521273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.521467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.521474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.521765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.521773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.521836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.521842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.521987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.521995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.522307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.522314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.522746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.522754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.522927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.522935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.523162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.523169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.523473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.523480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.523806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.523813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.524026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.524034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.524253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.524261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.524443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.524459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.524752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.524759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.525076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.525083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.525416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.525423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.525687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.525693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.525867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.525873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.526154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.526162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.526323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.526331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.526618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.526626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:47.975 [2024-11-26 19:27:00.526939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.526956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:30:47.975 [2024-11-26 19:27:00.527379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.527388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.527691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.527700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b9 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.975 0 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:47.975 [2024-11-26 19:27:00.527988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.527997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.528288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.528295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.528575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.528582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.528908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.528916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.529231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.529238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.529391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.529398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.529682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.529689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.529989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.529996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.530393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.530401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.975 [2024-11-26 19:27:00.530705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.975 [2024-11-26 19:27:00.530712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.975 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.531031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.531038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.531206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.531212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.531489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.531496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.531831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.531839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.532152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.532160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.532529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.532537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.532751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.532759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.533070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.533078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.533269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.533276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.533633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.533640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.533944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.533952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.534323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.534329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.534490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.534497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.534888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.534896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.535309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.535316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.535485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.535493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.535800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.535807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.536024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.536032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.536354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.536362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.536669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.536677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.536996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.537004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.537179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.537186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.537367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.537374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.537689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.537699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.538086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.538093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.538273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.538287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.538604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.538613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.538929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.538938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.539112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.539120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.539341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.539349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:47.976 [2024-11-26 19:27:00.539663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:47.976 [2024-11-26 19:27:00.539671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:47.976 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.539993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.540002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.540309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.540318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.540486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.540494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.540881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.540889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.541213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.541220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.541516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.541522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.541845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.541852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.542128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.542136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.542459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.542466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.542776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.542784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.542952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.542960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.543148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.543155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.543395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.543403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.543702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.543710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.544020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.544027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.544066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.544072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.544389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.544396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.544686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.544693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.545007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.545014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.545338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.545345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.545654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.545661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.545967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.545975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.546143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.245 [2024-11-26 19:27:00.546151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.245 qpair failed and we were unable to recover it. 00:30:48.245 [2024-11-26 19:27:00.546434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.546442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.546653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.546660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.546986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.546993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.547160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.547167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.547539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.547545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.547639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.547646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.547925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.547933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.548252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.548259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.548418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.548426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.548814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.548822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.549033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.549041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.549390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.549397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.549768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.549778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.550135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.550143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.550474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.550481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.550679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.550695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.551031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.551039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.551387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.551394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.551433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.551440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.551734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.551742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.552175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.552183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.552352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.552360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.552571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.552578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.552891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.552899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.553201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.553208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.553501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.553508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.553550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.553557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.553689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.553695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.554020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.554029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.554402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.554410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.554561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.554568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.554925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.554933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.555243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.555250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.555580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.555587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.555847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.555854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.556176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.556184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.556554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.556561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.556885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.246 [2024-11-26 19:27:00.556893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.246 qpair failed and we were unable to recover it. 00:30:48.246 [2024-11-26 19:27:00.557288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.557295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 Malloc0 00:30:48.247 [2024-11-26 19:27:00.557502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.557518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.557872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.557879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.558199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.558206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:48.247 [2024-11-26 19:27:00.558562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.558570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:30:48.247 [2024-11-26 19:27:00.558804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.558812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.558994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.559001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:48.247 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:48.247 [2024-11-26 19:27:00.559311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.559319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.559676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.559683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.560005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.560012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.560324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.560331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.560718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.560727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.560938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.560945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.561327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.561334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.561682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.561690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.561879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.561886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.562258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.562265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.562451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.562458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.562752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.562758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.562921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.562928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.562971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.562978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.563346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.563353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.563657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.563665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.564036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.564043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.564299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.564306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.564618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.564625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.564666] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.564673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.564957] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:48.247 [2024-11-26 19:27:00.565002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.565009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.565339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.565347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.565667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.565675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.566016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.566023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.566333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.566340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.566517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.566525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.566811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.566817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.567003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.567011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.247 qpair failed and we were unable to recover it. 00:30:48.247 [2024-11-26 19:27:00.567334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.247 [2024-11-26 19:27:00.567342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.567553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.567561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.567720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.567726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.567928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.567935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.567977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.567984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.568282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.568289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.568490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.568498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.568844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.568852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.569029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.569036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.569389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.569396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.569672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.569680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.569856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.569868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.570167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.570174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.570390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.570397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.570710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.570717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.570903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.570911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.571184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.571192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.571361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.571370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.571559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.571565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.571681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.571687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.571984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.571992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.572186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.572195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.572543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.572550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.572726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.572735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.572915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.572922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.573234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.573241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.573430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.573436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.573652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.573659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.573837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.573843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:48.248 [2024-11-26 19:27:00.574216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.574223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:30:48.248 [2024-11-26 19:27:00.574547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.574555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:48.248 [2024-11-26 19:27:00.574982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.574990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:48.248 [2024-11-26 19:27:00.575298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.575306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.575554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.575561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.575723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.575730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.576025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.248 [2024-11-26 19:27:00.576032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.248 qpair failed and we were unable to recover it. 00:30:48.248 [2024-11-26 19:27:00.576353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.576360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.576586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.576594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.576971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.576978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.577156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.577163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.577356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.577363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.577589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.577595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.577791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.577798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.578111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.578120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.578433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.578441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.578767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.578775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.578949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.578957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.579294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.579302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.579599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.579607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.579816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.579824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.580013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.580021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.580313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.580320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.580520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.580528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.580742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.580749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.581057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.581064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.581220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.581227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.581481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.581488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.581755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.581761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.582077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.582085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.582309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.582316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.582713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.582720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.582891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.582898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.583072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.583081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.583289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.583297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.583470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.583477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.583642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.583650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.583924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.583933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.584275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.584282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.584599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.249 [2024-11-26 19:27:00.584606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.249 qpair failed and we were unable to recover it. 00:30:48.249 [2024-11-26 19:27:00.584793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.584800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.585171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.585178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.585500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.585507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.585750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.585756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:48.250 [2024-11-26 19:27:00.586127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.586134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:30:48.250 [2024-11-26 19:27:00.586367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.586374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.586570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.586578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:48.250 [2024-11-26 19:27:00.586741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.586747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:48.250 [2024-11-26 19:27:00.586952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.586960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.587151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.587159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.587336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.587343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.587646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.587654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.588095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.588103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.588278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.588286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.588470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.588477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.588811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.588819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.589138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.589145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.589320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.589327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.589603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.589610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.589767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.589775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.590002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.590009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.590368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.590374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.590550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.590558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.590726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.590733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.590923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.590931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.591236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.591244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.591564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.591571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.591917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.591924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.592134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.592141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.592343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.592351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.592454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.592461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.592730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.592737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.593129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.593136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.593426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.593432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.593748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.593754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.250 [2024-11-26 19:27:00.593927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.250 [2024-11-26 19:27:00.593935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.250 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.594319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.594326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.594491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.594498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.594841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.594849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.595164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.595171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.595337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.595344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.595557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.595565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.595722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.595729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.595896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.595903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.596198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.596205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.596492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.596499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.596819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.596826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.597155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.597162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.597198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.597204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.597390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.597396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.597597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.597604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.597992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.598001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:48.251 [2024-11-26 19:27:00.598286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.598293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.598334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.598341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:30:48.251 [2024-11-26 19:27:00.598516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.598523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:48.251 [2024-11-26 19:27:00.598854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.598869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:48.251 [2024-11-26 19:27:00.599068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.599076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.599431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.599440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.599613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.599621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.599914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.599923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.600210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.600220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.600430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.600437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.600516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.600523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.600714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.600721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.601045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.601053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.601376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.601383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.601565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.601572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.601834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.601843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.602035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.602043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.602219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.602227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.602418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.602426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.602506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.251 [2024-11-26 19:27:00.602513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.251 qpair failed and we were unable to recover it. 00:30:48.251 [2024-11-26 19:27:00.602688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.602695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.602882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.602897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.603072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.603079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.603260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.603267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.603558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.603566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.603606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.603612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.603913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.603921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.604119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.604127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.604414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.604421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.604740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.604747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.605061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:30:48.252 [2024-11-26 19:27:00.605068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fb788000b90 with addr=10.0.0.2, port=4420 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.605247] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:48.252 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:48.252 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:30:48.252 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:48.252 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:48.252 [2024-11-26 19:27:00.615953] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.616019] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.616034] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.616039] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.616044] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.616061] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:48.252 19:27:00 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@50 -- # wait 4064962 00:30:48.252 [2024-11-26 19:27:00.625899] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.625958] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.625970] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.625975] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.625980] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.625991] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.635851] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.635904] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.635915] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.635919] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.635924] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.635934] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.645942] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.646036] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.646046] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.646051] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.646055] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.646065] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.655772] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.655820] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.655830] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.655835] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.655840] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.655850] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.665736] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.665812] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.665822] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.665830] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.665835] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.665845] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.675891] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.675972] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.675982] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.675987] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.675991] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.676001] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.685903] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.685958] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.685967] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.252 [2024-11-26 19:27:00.685972] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.252 [2024-11-26 19:27:00.685977] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.252 [2024-11-26 19:27:00.685987] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.252 qpair failed and we were unable to recover it. 00:30:48.252 [2024-11-26 19:27:00.695967] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.252 [2024-11-26 19:27:00.696017] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.252 [2024-11-26 19:27:00.696028] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.696033] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.696037] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.696047] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.705982] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.706033] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.706043] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.706048] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.706053] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.706066] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.716005] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.716048] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.716057] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.716062] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.716067] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.716077] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.726011] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.726061] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.726071] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.726076] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.726081] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.726091] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.736087] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.736140] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.736151] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.736157] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.736161] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.736172] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.746092] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.746148] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.746159] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.746164] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.746168] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.746179] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.755974] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.756024] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.756035] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.756040] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.756045] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.756055] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.766161] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.766209] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.766219] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.766224] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.766229] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.766239] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.776046] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.776102] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.776112] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.776117] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.776121] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.776132] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.786168] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.786218] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.786227] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.786233] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.786238] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.786248] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.796234] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.796280] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.796292] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.796297] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.796302] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.796312] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.806122] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.806171] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.806180] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.806185] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.806189] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.806199] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.816305] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.816355] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.816365] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.816369] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.816374] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.253 [2024-11-26 19:27:00.816384] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.253 qpair failed and we were unable to recover it. 00:30:48.253 [2024-11-26 19:27:00.826326] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.253 [2024-11-26 19:27:00.826374] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.253 [2024-11-26 19:27:00.826384] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.253 [2024-11-26 19:27:00.826389] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.253 [2024-11-26 19:27:00.826393] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.254 [2024-11-26 19:27:00.826403] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.254 qpair failed and we were unable to recover it. 00:30:48.254 [2024-11-26 19:27:00.836329] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.254 [2024-11-26 19:27:00.836382] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.254 [2024-11-26 19:27:00.836392] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.254 [2024-11-26 19:27:00.836397] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.254 [2024-11-26 19:27:00.836405] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.254 [2024-11-26 19:27:00.836415] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.254 qpair failed and we were unable to recover it. 00:30:48.254 [2024-11-26 19:27:00.846484] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.254 [2024-11-26 19:27:00.846576] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.254 [2024-11-26 19:27:00.846586] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.254 [2024-11-26 19:27:00.846590] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.254 [2024-11-26 19:27:00.846595] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.254 [2024-11-26 19:27:00.846605] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.254 qpair failed and we were unable to recover it. 00:30:48.254 [2024-11-26 19:27:00.856449] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.254 [2024-11-26 19:27:00.856499] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.254 [2024-11-26 19:27:00.856509] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.254 [2024-11-26 19:27:00.856514] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.254 [2024-11-26 19:27:00.856518] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.254 [2024-11-26 19:27:00.856528] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.254 qpair failed and we were unable to recover it. 00:30:48.516 [2024-11-26 19:27:00.866467] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.516 [2024-11-26 19:27:00.866515] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.516 [2024-11-26 19:27:00.866525] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.516 [2024-11-26 19:27:00.866530] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.516 [2024-11-26 19:27:00.866534] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.516 [2024-11-26 19:27:00.866544] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.516 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.876488] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.876530] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.876540] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.876545] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.876549] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.876559] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.886449] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.886529] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.886539] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.886544] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.886548] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.886558] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.896506] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.896556] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.896565] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.896570] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.896575] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.896584] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.906534] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.906585] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.906604] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.906610] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.906615] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.906629] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.916612] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.916667] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.916685] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.916691] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.916696] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.916710] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.926588] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.926641] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.926666] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.926672] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.926677] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.926691] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.936618] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.936674] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.936693] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.936699] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.936704] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.936718] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.946621] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.946683] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.946694] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.946699] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.946704] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.946715] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.956666] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.956716] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.956727] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.956732] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.956737] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.956747] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.966650] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.966698] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.966708] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.966713] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.517 [2024-11-26 19:27:00.966721] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.517 [2024-11-26 19:27:00.966731] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.517 qpair failed and we were unable to recover it. 00:30:48.517 [2024-11-26 19:27:00.976727] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.517 [2024-11-26 19:27:00.976773] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.517 [2024-11-26 19:27:00.976783] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.517 [2024-11-26 19:27:00.976788] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:00.976792] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:00.976802] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:00.986759] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:00.986849] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:00.986859] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:00.986868] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:00.986872] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:00.986882] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:00.996629] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:00.996676] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:00.996686] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:00.996691] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:00.996695] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:00.996705] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.006795] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.006845] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.006855] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.006859] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.006869] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.006879] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.016880] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.016948] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.016957] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.016963] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.016967] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.016977] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.026859] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.026915] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.026925] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.026930] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.026934] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.026944] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.036753] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.036801] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.036811] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.036816] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.036820] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.036830] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.046785] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.046837] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.046846] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.046851] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.046856] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.046871] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.056913] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.056977] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.056987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.056992] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.056997] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.057007] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.066867] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.066919] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.066928] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.066933] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.066938] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.066948] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.518 qpair failed and we were unable to recover it. 00:30:48.518 [2024-11-26 19:27:01.077010] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.518 [2024-11-26 19:27:01.077091] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.518 [2024-11-26 19:27:01.077102] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.518 [2024-11-26 19:27:01.077106] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.518 [2024-11-26 19:27:01.077111] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.518 [2024-11-26 19:27:01.077121] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.519 qpair failed and we were unable to recover it. 00:30:48.519 [2024-11-26 19:27:01.087003] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.519 [2024-11-26 19:27:01.087056] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.519 [2024-11-26 19:27:01.087065] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.519 [2024-11-26 19:27:01.087070] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.519 [2024-11-26 19:27:01.087074] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.519 [2024-11-26 19:27:01.087084] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.519 qpair failed and we were unable to recover it. 00:30:48.519 [2024-11-26 19:27:01.097070] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.519 [2024-11-26 19:27:01.097120] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.519 [2024-11-26 19:27:01.097129] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.519 [2024-11-26 19:27:01.097137] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.519 [2024-11-26 19:27:01.097142] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.519 [2024-11-26 19:27:01.097152] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.519 qpair failed and we were unable to recover it. 00:30:48.519 [2024-11-26 19:27:01.107085] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.519 [2024-11-26 19:27:01.107131] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.519 [2024-11-26 19:27:01.107140] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.519 [2024-11-26 19:27:01.107145] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.519 [2024-11-26 19:27:01.107149] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.519 [2024-11-26 19:27:01.107159] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.519 qpair failed and we were unable to recover it. 00:30:48.519 [2024-11-26 19:27:01.116981] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.519 [2024-11-26 19:27:01.117033] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.519 [2024-11-26 19:27:01.117043] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.519 [2024-11-26 19:27:01.117047] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.519 [2024-11-26 19:27:01.117052] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.519 [2024-11-26 19:27:01.117062] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.519 qpair failed and we were unable to recover it. 00:30:48.519 [2024-11-26 19:27:01.127133] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.519 [2024-11-26 19:27:01.127186] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.519 [2024-11-26 19:27:01.127195] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.519 [2024-11-26 19:27:01.127200] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.519 [2024-11-26 19:27:01.127205] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.519 [2024-11-26 19:27:01.127215] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.519 qpair failed and we were unable to recover it. 00:30:48.519 [2024-11-26 19:27:01.137176] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.519 [2024-11-26 19:27:01.137224] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.519 [2024-11-26 19:27:01.137234] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.519 [2024-11-26 19:27:01.137239] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.519 [2024-11-26 19:27:01.137243] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.519 [2024-11-26 19:27:01.137256] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.519 qpair failed and we were unable to recover it. 00:30:48.781 [2024-11-26 19:27:01.147191] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.781 [2024-11-26 19:27:01.147244] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.781 [2024-11-26 19:27:01.147254] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.781 [2024-11-26 19:27:01.147259] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.781 [2024-11-26 19:27:01.147263] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.781 [2024-11-26 19:27:01.147273] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.781 qpair failed and we were unable to recover it. 00:30:48.781 [2024-11-26 19:27:01.157191] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.781 [2024-11-26 19:27:01.157243] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.781 [2024-11-26 19:27:01.157252] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.781 [2024-11-26 19:27:01.157257] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.157262] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.157272] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.167227] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.167278] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.167288] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.167293] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.167297] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.167307] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.177298] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.177400] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.177410] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.177415] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.177420] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.177430] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.187279] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.187333] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.187343] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.187348] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.187352] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.187363] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.197316] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.197367] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.197377] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.197382] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.197386] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.197397] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.207345] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.207395] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.207405] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.207410] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.207414] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.207425] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.217374] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.217466] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.217476] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.217480] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.217485] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.217495] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.227364] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.227410] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.227419] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.227427] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.227431] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.227441] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.237416] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.237462] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.237472] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.237477] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.237481] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.237491] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.247316] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.782 [2024-11-26 19:27:01.247412] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.782 [2024-11-26 19:27:01.247422] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.782 [2024-11-26 19:27:01.247427] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.782 [2024-11-26 19:27:01.247432] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.782 [2024-11-26 19:27:01.247442] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.782 qpair failed and we were unable to recover it. 00:30:48.782 [2024-11-26 19:27:01.257352] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.257416] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.257427] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.257432] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.257436] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.257447] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.267496] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.267547] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.267557] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.267562] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.267567] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.267580] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.277525] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.277576] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.277586] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.277591] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.277595] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.277605] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.287564] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.287613] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.287623] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.287628] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.287632] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.287642] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.297576] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.297629] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.297639] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.297644] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.297648] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.297658] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.307601] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.307650] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.307659] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.307664] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.307668] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.307679] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.317610] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.317654] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.317663] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.317668] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.317673] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.317682] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.327678] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.327729] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.327738] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.327743] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.327748] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.327757] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.337715] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.337766] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.337775] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.337780] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.337785] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.337795] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.347702] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.347748] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.783 [2024-11-26 19:27:01.347758] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.783 [2024-11-26 19:27:01.347763] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.783 [2024-11-26 19:27:01.347767] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.783 [2024-11-26 19:27:01.347777] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.783 qpair failed and we were unable to recover it. 00:30:48.783 [2024-11-26 19:27:01.357795] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.783 [2024-11-26 19:27:01.357874] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.784 [2024-11-26 19:27:01.357888] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.784 [2024-11-26 19:27:01.357893] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.784 [2024-11-26 19:27:01.357897] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.784 [2024-11-26 19:27:01.357907] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.784 qpair failed and we were unable to recover it. 00:30:48.784 [2024-11-26 19:27:01.367789] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.784 [2024-11-26 19:27:01.367844] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.784 [2024-11-26 19:27:01.367854] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.784 [2024-11-26 19:27:01.367859] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.784 [2024-11-26 19:27:01.367867] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.784 [2024-11-26 19:27:01.367878] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.784 qpair failed and we were unable to recover it. 00:30:48.784 [2024-11-26 19:27:01.377720] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.784 [2024-11-26 19:27:01.377773] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.784 [2024-11-26 19:27:01.377782] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.784 [2024-11-26 19:27:01.377787] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.784 [2024-11-26 19:27:01.377792] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.784 [2024-11-26 19:27:01.377802] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.784 qpair failed and we were unable to recover it. 00:30:48.784 [2024-11-26 19:27:01.387841] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.784 [2024-11-26 19:27:01.387927] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.784 [2024-11-26 19:27:01.387936] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.784 [2024-11-26 19:27:01.387941] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.784 [2024-11-26 19:27:01.387945] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.784 [2024-11-26 19:27:01.387956] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.784 qpair failed and we were unable to recover it. 00:30:48.784 [2024-11-26 19:27:01.397876] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:48.784 [2024-11-26 19:27:01.397926] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:48.784 [2024-11-26 19:27:01.397935] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:48.784 [2024-11-26 19:27:01.397940] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:48.784 [2024-11-26 19:27:01.397947] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:48.784 [2024-11-26 19:27:01.397958] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:48.784 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.407915] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.407972] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.407981] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.407986] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.407991] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.408001] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.417932] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.417983] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.417993] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.417998] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.418002] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.418012] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.427960] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.428045] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.428055] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.428059] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.428064] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.428074] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.437959] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.438020] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.438029] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.438034] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.438038] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.438048] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.448018] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.448067] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.448077] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.448082] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.448087] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.448097] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.458047] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.458121] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.458131] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.458136] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.458141] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.458151] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.468075] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.468149] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.468158] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.468163] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.468167] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.468177] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.478074] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.478126] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.478136] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.478141] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.478145] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.478155] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.488159] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.488208] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.488220] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.488225] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.488229] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.488239] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.047 [2024-11-26 19:27:01.498139] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.047 [2024-11-26 19:27:01.498220] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.047 [2024-11-26 19:27:01.498230] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.047 [2024-11-26 19:27:01.498235] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.047 [2024-11-26 19:27:01.498239] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.047 [2024-11-26 19:27:01.498249] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.047 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.508136] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.508182] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.508192] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.508197] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.508201] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.508211] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.518170] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.518216] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.518226] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.518231] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.518235] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.518245] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.528278] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.528377] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.528386] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.528391] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.528398] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.528409] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.538277] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.538328] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.538338] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.538343] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.538347] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.538357] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.548289] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.548338] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.548347] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.548352] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.548356] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.548366] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.558322] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.558369] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.558378] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.558383] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.558388] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.558397] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.568352] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.568403] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.568413] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.568418] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.568422] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.568432] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.578347] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.578434] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.578444] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.578449] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.578453] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.578463] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.588267] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.588314] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.588324] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.588329] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.588333] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.588343] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.048 [2024-11-26 19:27:01.598423] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.048 [2024-11-26 19:27:01.598469] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.048 [2024-11-26 19:27:01.598479] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.048 [2024-11-26 19:27:01.598484] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.048 [2024-11-26 19:27:01.598488] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.048 [2024-11-26 19:27:01.598498] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.048 qpair failed and we were unable to recover it. 00:30:49.049 [2024-11-26 19:27:01.608502] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.049 [2024-11-26 19:27:01.608557] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.049 [2024-11-26 19:27:01.608567] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.049 [2024-11-26 19:27:01.608571] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.049 [2024-11-26 19:27:01.608576] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.049 [2024-11-26 19:27:01.608586] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.049 qpair failed and we were unable to recover it. 00:30:49.049 [2024-11-26 19:27:01.618494] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.049 [2024-11-26 19:27:01.618548] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.049 [2024-11-26 19:27:01.618558] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.049 [2024-11-26 19:27:01.618563] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.049 [2024-11-26 19:27:01.618567] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.049 [2024-11-26 19:27:01.618577] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.049 qpair failed and we were unable to recover it. 00:30:49.049 [2024-11-26 19:27:01.628506] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.049 [2024-11-26 19:27:01.628576] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.049 [2024-11-26 19:27:01.628586] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.049 [2024-11-26 19:27:01.628591] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.049 [2024-11-26 19:27:01.628596] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.049 [2024-11-26 19:27:01.628606] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.049 qpair failed and we were unable to recover it. 00:30:49.049 [2024-11-26 19:27:01.638532] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.049 [2024-11-26 19:27:01.638580] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.049 [2024-11-26 19:27:01.638599] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.049 [2024-11-26 19:27:01.638605] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.049 [2024-11-26 19:27:01.638610] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.049 [2024-11-26 19:27:01.638625] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.049 qpair failed and we were unable to recover it. 00:30:49.049 [2024-11-26 19:27:01.648575] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.049 [2024-11-26 19:27:01.648631] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.049 [2024-11-26 19:27:01.648649] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.049 [2024-11-26 19:27:01.648655] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.049 [2024-11-26 19:27:01.648660] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.049 [2024-11-26 19:27:01.648674] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.049 qpair failed and we were unable to recover it. 00:30:49.049 [2024-11-26 19:27:01.658610] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.049 [2024-11-26 19:27:01.658706] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.049 [2024-11-26 19:27:01.658717] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.049 [2024-11-26 19:27:01.658726] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.049 [2024-11-26 19:27:01.658731] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.049 [2024-11-26 19:27:01.658742] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.049 qpair failed and we were unable to recover it. 00:30:49.320 [2024-11-26 19:27:01.668614] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.320 [2024-11-26 19:27:01.668658] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.320 [2024-11-26 19:27:01.668668] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.320 [2024-11-26 19:27:01.668673] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.320 [2024-11-26 19:27:01.668678] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.320 [2024-11-26 19:27:01.668689] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.320 qpair failed and we were unable to recover it. 00:30:49.320 [2024-11-26 19:27:01.678629] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.320 [2024-11-26 19:27:01.678674] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.320 [2024-11-26 19:27:01.678684] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.320 [2024-11-26 19:27:01.678689] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.320 [2024-11-26 19:27:01.678693] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.320 [2024-11-26 19:27:01.678704] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.320 qpair failed and we were unable to recover it. 00:30:49.320 [2024-11-26 19:27:01.688672] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.320 [2024-11-26 19:27:01.688723] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.320 [2024-11-26 19:27:01.688732] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.320 [2024-11-26 19:27:01.688737] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.320 [2024-11-26 19:27:01.688741] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.320 [2024-11-26 19:27:01.688751] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.320 qpair failed and we were unable to recover it. 00:30:49.320 [2024-11-26 19:27:01.698709] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.320 [2024-11-26 19:27:01.698765] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.320 [2024-11-26 19:27:01.698775] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.320 [2024-11-26 19:27:01.698779] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.320 [2024-11-26 19:27:01.698784] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.320 [2024-11-26 19:27:01.698801] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.320 qpair failed and we were unable to recover it. 00:30:49.320 [2024-11-26 19:27:01.708614] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.320 [2024-11-26 19:27:01.708659] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.320 [2024-11-26 19:27:01.708669] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.320 [2024-11-26 19:27:01.708674] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.320 [2024-11-26 19:27:01.708678] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.320 [2024-11-26 19:27:01.708688] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.320 qpair failed and we were unable to recover it. 00:30:49.320 [2024-11-26 19:27:01.718766] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.320 [2024-11-26 19:27:01.718816] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.320 [2024-11-26 19:27:01.718826] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.320 [2024-11-26 19:27:01.718831] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.320 [2024-11-26 19:27:01.718836] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.320 [2024-11-26 19:27:01.718846] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.320 qpair failed and we were unable to recover it. 00:30:49.320 [2024-11-26 19:27:01.728797] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.320 [2024-11-26 19:27:01.728846] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.728855] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.728860] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.728869] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.728879] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.738831] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.738879] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.738889] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.738894] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.738898] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.738908] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.748836] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.748919] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.748929] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.748934] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.748938] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.748948] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.758805] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.758851] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.758873] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.758878] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.758883] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.758897] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.768891] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.768940] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.768950] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.768955] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.768959] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.768970] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.778926] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.778977] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.778988] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.778993] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.778997] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.779007] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.788933] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.788980] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.788992] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.788997] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.789002] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.789012] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.798965] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.799014] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.799024] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.799029] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.799033] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.799043] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.808883] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.808937] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.808947] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.808952] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.808956] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.808967] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.819027] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.819077] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.819087] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.819092] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.819096] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.819107] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.829025] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.829074] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.829084] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.829089] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.829094] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.829106] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.839088] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.839139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.839149] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.839154] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.839158] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.839168] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.849092] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.849145] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.849154] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.849159] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.849164] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.849174] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.859152] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.859207] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.859217] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.859221] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.859226] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.859236] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.869124] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.869170] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.869180] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.869185] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.869189] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.869199] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.879064] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.879120] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.321 [2024-11-26 19:27:01.879130] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.321 [2024-11-26 19:27:01.879135] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.321 [2024-11-26 19:27:01.879140] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.321 [2024-11-26 19:27:01.879150] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.321 qpair failed and we were unable to recover it. 00:30:49.321 [2024-11-26 19:27:01.889245] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.321 [2024-11-26 19:27:01.889296] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.322 [2024-11-26 19:27:01.889306] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.322 [2024-11-26 19:27:01.889311] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.322 [2024-11-26 19:27:01.889316] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.322 [2024-11-26 19:27:01.889326] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.322 qpair failed and we were unable to recover it. 00:30:49.322 [2024-11-26 19:27:01.899247] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.322 [2024-11-26 19:27:01.899297] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.322 [2024-11-26 19:27:01.899307] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.322 [2024-11-26 19:27:01.899312] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.322 [2024-11-26 19:27:01.899317] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.322 [2024-11-26 19:27:01.899327] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.322 qpair failed and we were unable to recover it. 00:30:49.322 [2024-11-26 19:27:01.909147] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.322 [2024-11-26 19:27:01.909195] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.322 [2024-11-26 19:27:01.909206] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.322 [2024-11-26 19:27:01.909211] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.322 [2024-11-26 19:27:01.909216] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.322 [2024-11-26 19:27:01.909227] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.322 qpair failed and we were unable to recover it. 00:30:49.322 [2024-11-26 19:27:01.919175] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.322 [2024-11-26 19:27:01.919229] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.322 [2024-11-26 19:27:01.919243] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.322 [2024-11-26 19:27:01.919248] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.322 [2024-11-26 19:27:01.919253] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.322 [2024-11-26 19:27:01.919263] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.322 qpair failed and we were unable to recover it. 00:30:49.322 [2024-11-26 19:27:01.929202] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.322 [2024-11-26 19:27:01.929262] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.322 [2024-11-26 19:27:01.929271] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.322 [2024-11-26 19:27:01.929277] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.322 [2024-11-26 19:27:01.929281] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.322 [2024-11-26 19:27:01.929291] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.322 qpair failed and we were unable to recover it. 00:30:49.322 [2024-11-26 19:27:01.939367] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.322 [2024-11-26 19:27:01.939414] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.322 [2024-11-26 19:27:01.939424] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.322 [2024-11-26 19:27:01.939429] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.322 [2024-11-26 19:27:01.939434] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.322 [2024-11-26 19:27:01.939444] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.322 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:01.949369] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:01.949417] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:01.949428] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:01.949433] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:01.949437] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:01.949447] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:01.959406] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:01.959454] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:01.959464] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:01.959470] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:01.959477] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:01.959487] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:01.969421] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:01.969473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:01.969483] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:01.969488] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:01.969492] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:01.969502] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:01.979473] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:01.979521] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:01.979531] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:01.979537] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:01.979541] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:01.979551] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:01.989457] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:01.989508] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:01.989517] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:01.989523] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:01.989527] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:01.989537] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:01.999503] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:01.999556] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:01.999566] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:01.999571] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:01.999576] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:01.999586] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:02.009534] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:02.009583] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:02.009593] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:02.009598] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:02.009603] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:02.009613] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:02.019585] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:02.019676] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:02.019685] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:02.019692] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:02.019696] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:02.019706] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:02.029578] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:02.029658] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:02.029667] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:02.029672] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:02.029677] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:02.029689] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:02.039620] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:02.039669] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:02.039679] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:02.039685] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:02.039690] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:02.039700] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:02.049660] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.583 [2024-11-26 19:27:02.049709] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.583 [2024-11-26 19:27:02.049722] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.583 [2024-11-26 19:27:02.049727] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.583 [2024-11-26 19:27:02.049732] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.583 [2024-11-26 19:27:02.049742] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.583 qpair failed and we were unable to recover it. 00:30:49.583 [2024-11-26 19:27:02.059688] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.059763] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.059773] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.059779] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.059783] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.059794] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.069716] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.069758] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.069768] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.069774] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.069779] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.069789] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.079731] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.079783] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.079793] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.079798] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.079803] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.079814] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.089734] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.089801] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.089811] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.089819] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.089824] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.089835] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.099806] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.099854] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.099869] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.099875] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.099880] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.099891] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.109825] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.109879] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.109889] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.109895] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.109899] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.109910] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.119848] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.119903] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.119913] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.119918] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.119923] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.119933] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.129878] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.129966] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.129975] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.129981] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.129986] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.129997] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.139890] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.139939] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.139949] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.139954] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.139959] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.139969] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.149799] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.149848] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.149857] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.149867] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.149872] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.149882] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.159842] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.159902] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.159912] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.159918] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.159922] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.159933] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.169968] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.170017] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.170027] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.170032] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.170037] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.170047] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.180053] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.180139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.180149] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.584 [2024-11-26 19:27:02.180154] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.584 [2024-11-26 19:27:02.180159] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.584 [2024-11-26 19:27:02.180170] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.584 qpair failed and we were unable to recover it. 00:30:49.584 [2024-11-26 19:27:02.189910] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.584 [2024-11-26 19:27:02.189955] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.584 [2024-11-26 19:27:02.189965] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.585 [2024-11-26 19:27:02.189970] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.585 [2024-11-26 19:27:02.189975] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.585 [2024-11-26 19:27:02.189985] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.585 qpair failed and we were unable to recover it. 00:30:49.585 [2024-11-26 19:27:02.199942] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.585 [2024-11-26 19:27:02.199992] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.585 [2024-11-26 19:27:02.200001] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.585 [2024-11-26 19:27:02.200007] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.585 [2024-11-26 19:27:02.200012] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.585 [2024-11-26 19:27:02.200023] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.585 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.210092] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.210147] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.210157] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.210162] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.210167] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.210178] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.220204] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.220258] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.220267] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.220276] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.220281] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.220291] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.230168] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.230219] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.230229] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.230234] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.230239] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.230249] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.240170] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.240216] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.240226] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.240231] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.240236] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.240246] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.250217] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.250274] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.250284] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.250289] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.250294] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.250304] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.260247] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.260295] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.260305] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.260311] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.260316] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.260329] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.270301] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.270345] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.270354] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.270359] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.270364] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.270374] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.280307] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.280380] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.280390] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.280396] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.280400] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.280410] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.290344] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.290395] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.290405] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.290410] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.290414] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.290424] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.300375] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.300425] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.300434] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.300439] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.300444] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.847 [2024-11-26 19:27:02.300454] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.847 qpair failed and we were unable to recover it. 00:30:49.847 [2024-11-26 19:27:02.310361] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.847 [2024-11-26 19:27:02.310413] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.847 [2024-11-26 19:27:02.310423] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.847 [2024-11-26 19:27:02.310428] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.847 [2024-11-26 19:27:02.310433] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.310443] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.320376] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.320419] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.320429] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.320436] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.320441] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.320451] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.330450] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.330539] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.330549] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.330554] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.330559] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.330570] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.340466] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.340515] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.340525] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.340530] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.340535] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.340545] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.350505] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.350555] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.350567] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.350573] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.350578] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.350588] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.360387] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.360432] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.360442] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.360447] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.360452] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.360462] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.370536] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.370589] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.370598] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.370604] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.370609] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.370619] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.380591] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.380654] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.380673] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.380679] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.380684] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.380699] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.390593] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.390686] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.390705] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.390712] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.390717] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.390738] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.400610] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.400699] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.400711] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.400717] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.400722] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.400736] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.410672] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.410721] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.410732] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.410737] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.410742] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.410753] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.420614] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.420664] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.420674] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.420679] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.420683] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.420694] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.430580] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.430633] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.430643] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.430648] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.430653] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.430664] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.848 [2024-11-26 19:27:02.440744] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.848 [2024-11-26 19:27:02.440797] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.848 [2024-11-26 19:27:02.440808] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.848 [2024-11-26 19:27:02.440816] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.848 [2024-11-26 19:27:02.440822] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.848 [2024-11-26 19:27:02.440834] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.848 qpair failed and we were unable to recover it. 00:30:49.849 [2024-11-26 19:27:02.450667] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.849 [2024-11-26 19:27:02.450727] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.849 [2024-11-26 19:27:02.450738] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.849 [2024-11-26 19:27:02.450744] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.849 [2024-11-26 19:27:02.450749] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.849 [2024-11-26 19:27:02.450760] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.849 qpair failed and we were unable to recover it. 00:30:49.849 [2024-11-26 19:27:02.460819] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:49.849 [2024-11-26 19:27:02.460872] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:49.849 [2024-11-26 19:27:02.460883] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:49.849 [2024-11-26 19:27:02.460888] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:49.849 [2024-11-26 19:27:02.460893] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:49.849 [2024-11-26 19:27:02.460904] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:49.849 qpair failed and we were unable to recover it. 00:30:50.221 [2024-11-26 19:27:02.470819] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.221 [2024-11-26 19:27:02.470869] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.221 [2024-11-26 19:27:02.470879] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.221 [2024-11-26 19:27:02.470885] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.221 [2024-11-26 19:27:02.470889] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.221 [2024-11-26 19:27:02.470900] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.221 qpair failed and we were unable to recover it. 00:30:50.221 [2024-11-26 19:27:02.480904] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.221 [2024-11-26 19:27:02.480969] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.480982] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.480987] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.480992] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.481002] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.490897] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.490947] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.490957] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.490962] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.490967] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.490977] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.500909] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.500962] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.500972] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.500977] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.500981] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.500992] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.510907] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.510968] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.510978] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.510983] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.510988] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.510998] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.520952] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.521005] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.521015] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.521020] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.521028] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.521039] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.530967] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.531022] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.531032] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.531037] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.531042] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.531052] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.541003] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.541057] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.541067] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.541072] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.541077] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.541087] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.551031] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.551079] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.551089] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.551094] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.551099] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.551110] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.561078] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.561127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.561137] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.561143] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.561147] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.561158] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.571121] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.571172] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.571182] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.571188] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.571193] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.571203] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.581129] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.581201] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.581211] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.581216] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.581222] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.581233] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.591148] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.591192] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.591202] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.591208] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.591212] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.591223] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.601183] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.601230] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.601240] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.601245] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.601250] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.601261] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.611214] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.611264] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.611276] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.611281] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.611286] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.611296] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.222 qpair failed and we were unable to recover it. 00:30:50.222 [2024-11-26 19:27:02.621253] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.222 [2024-11-26 19:27:02.621323] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.222 [2024-11-26 19:27:02.621333] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.222 [2024-11-26 19:27:02.621338] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.222 [2024-11-26 19:27:02.621343] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.222 [2024-11-26 19:27:02.621353] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.631236] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.631284] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.631293] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.631299] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.631303] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.631314] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.641262] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.641346] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.641356] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.641361] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.641366] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.641377] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.651322] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.651369] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.651378] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.651387] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.651392] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.651402] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.661292] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.661344] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.661354] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.661359] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.661364] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.661374] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.671389] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.671474] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.671484] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.671489] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.671495] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.671505] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.681404] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.681457] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.681466] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.681471] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.681476] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.681486] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.691454] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.691506] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.691516] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.691521] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.691526] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.691537] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.701499] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.701550] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.701560] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.701565] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.701569] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.701579] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.711374] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.711421] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.711431] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.711436] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.711441] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.711451] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.721424] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.721482] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.721492] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.721497] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.721501] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.721511] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.731444] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.731497] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.731507] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.731513] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.731517] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.731528] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.741473] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.741529] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.741539] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.741544] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.741549] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.741560] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.751623] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.751675] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.751685] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.751690] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.751695] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.751706] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.761643] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.761689] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.761699] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.223 [2024-11-26 19:27:02.761705] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.223 [2024-11-26 19:27:02.761710] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.223 [2024-11-26 19:27:02.761720] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.223 qpair failed and we were unable to recover it. 00:30:50.223 [2024-11-26 19:27:02.771692] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.223 [2024-11-26 19:27:02.771748] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.223 [2024-11-26 19:27:02.771758] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.771763] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.771768] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.771778] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.224 [2024-11-26 19:27:02.781719] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.224 [2024-11-26 19:27:02.781771] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.224 [2024-11-26 19:27:02.781780] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.781788] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.781793] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.781803] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.224 [2024-11-26 19:27:02.791736] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.224 [2024-11-26 19:27:02.791782] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.224 [2024-11-26 19:27:02.791793] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.791798] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.791803] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.791814] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.224 [2024-11-26 19:27:02.801756] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.224 [2024-11-26 19:27:02.801802] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.224 [2024-11-26 19:27:02.801812] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.801818] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.801822] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.801833] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.224 [2024-11-26 19:27:02.811783] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.224 [2024-11-26 19:27:02.811833] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.224 [2024-11-26 19:27:02.811843] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.811848] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.811853] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.811868] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.224 [2024-11-26 19:27:02.821765] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.224 [2024-11-26 19:27:02.821817] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.224 [2024-11-26 19:27:02.821827] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.821833] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.821838] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.821851] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.224 [2024-11-26 19:27:02.831857] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.224 [2024-11-26 19:27:02.831908] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.224 [2024-11-26 19:27:02.831919] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.831924] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.831930] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.831941] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.224 [2024-11-26 19:27:02.841851] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.224 [2024-11-26 19:27:02.841917] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.224 [2024-11-26 19:27:02.841927] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.224 [2024-11-26 19:27:02.841932] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.224 [2024-11-26 19:27:02.841937] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.224 [2024-11-26 19:27:02.841948] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.224 qpair failed and we were unable to recover it. 00:30:50.485 [2024-11-26 19:27:02.851890] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.485 [2024-11-26 19:27:02.851947] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.485 [2024-11-26 19:27:02.851958] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.485 [2024-11-26 19:27:02.851963] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.485 [2024-11-26 19:27:02.851968] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.485 [2024-11-26 19:27:02.851978] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.485 qpair failed and we were unable to recover it. 00:30:50.485 [2024-11-26 19:27:02.861962] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.485 [2024-11-26 19:27:02.862026] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.485 [2024-11-26 19:27:02.862037] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.485 [2024-11-26 19:27:02.862042] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.485 [2024-11-26 19:27:02.862047] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.485 [2024-11-26 19:27:02.862057] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.485 qpair failed and we were unable to recover it. 00:30:50.485 [2024-11-26 19:27:02.872002] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.485 [2024-11-26 19:27:02.872051] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.485 [2024-11-26 19:27:02.872061] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.485 [2024-11-26 19:27:02.872066] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.872071] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.872082] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.882028] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.882077] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.882087] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.882092] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.882097] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.882107] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.892015] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.892069] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.892079] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.892085] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.892090] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.892101] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.902049] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.902100] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.902110] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.902115] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.902120] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.902130] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.912100] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.912183] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.912195] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.912200] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.912206] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.912216] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.921986] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.922030] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.922040] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.922046] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.922050] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.922061] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.932138] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.932190] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.932200] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.932205] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.932210] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.932220] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.942186] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.942235] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.942244] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.942250] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.942255] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.942265] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.952069] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.952127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.952136] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.952142] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.952149] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.952160] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.962180] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.962235] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.962245] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.962250] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.962255] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.962265] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.972226] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.972325] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.972335] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.972340] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.972345] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.972355] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.982156] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.982207] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.982217] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.982223] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.982227] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.982238] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.486 qpair failed and we were unable to recover it. 00:30:50.486 [2024-11-26 19:27:02.992170] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.486 [2024-11-26 19:27:02.992216] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.486 [2024-11-26 19:27:02.992226] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.486 [2024-11-26 19:27:02.992231] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.486 [2024-11-26 19:27:02.992236] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.486 [2024-11-26 19:27:02.992247] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.002321] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.002374] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.002384] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.002389] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.002394] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.002404] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.012335] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.012386] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.012396] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.012401] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.012406] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.012416] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.022395] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.022481] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.022491] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.022496] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.022501] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.022512] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.032421] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.032466] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.032475] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.032481] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.032486] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.032496] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.042485] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.042551] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.042563] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.042568] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.042573] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.042584] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.052484] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.052531] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.052541] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.052546] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.052551] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.052561] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.062524] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.062575] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.062585] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.062591] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.062595] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.062606] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.072540] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.072591] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.072601] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.072606] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.072611] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.072621] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.082573] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.082665] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.082674] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.082681] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.082691] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.082702] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.487 [2024-11-26 19:27:03.092596] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.487 [2024-11-26 19:27:03.092648] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.487 [2024-11-26 19:27:03.092658] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.487 [2024-11-26 19:27:03.092664] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.487 [2024-11-26 19:27:03.092668] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.487 [2024-11-26 19:27:03.092678] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.487 qpair failed and we were unable to recover it. 00:30:50.488 [2024-11-26 19:27:03.102633] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.488 [2024-11-26 19:27:03.102690] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.488 [2024-11-26 19:27:03.102700] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.488 [2024-11-26 19:27:03.102705] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.488 [2024-11-26 19:27:03.102710] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.488 [2024-11-26 19:27:03.102720] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.488 qpair failed and we were unable to recover it. 00:30:50.749 [2024-11-26 19:27:03.112610] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.749 [2024-11-26 19:27:03.112663] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.749 [2024-11-26 19:27:03.112673] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.749 [2024-11-26 19:27:03.112678] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.749 [2024-11-26 19:27:03.112683] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.749 [2024-11-26 19:27:03.112693] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.749 qpair failed and we were unable to recover it. 00:30:50.749 [2024-11-26 19:27:03.122673] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.749 [2024-11-26 19:27:03.122741] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.749 [2024-11-26 19:27:03.122752] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.749 [2024-11-26 19:27:03.122757] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.749 [2024-11-26 19:27:03.122762] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.749 [2024-11-26 19:27:03.122772] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.749 qpair failed and we were unable to recover it. 00:30:50.749 [2024-11-26 19:27:03.132575] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.749 [2024-11-26 19:27:03.132669] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.749 [2024-11-26 19:27:03.132678] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.749 [2024-11-26 19:27:03.132684] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.749 [2024-11-26 19:27:03.132689] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.749 [2024-11-26 19:27:03.132699] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.749 qpair failed and we were unable to recover it. 00:30:50.749 [2024-11-26 19:27:03.142746] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.749 [2024-11-26 19:27:03.142797] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.749 [2024-11-26 19:27:03.142807] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.749 [2024-11-26 19:27:03.142812] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.749 [2024-11-26 19:27:03.142817] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.749 [2024-11-26 19:27:03.142827] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.749 qpair failed and we were unable to recover it. 00:30:50.749 [2024-11-26 19:27:03.152700] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.749 [2024-11-26 19:27:03.152749] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.749 [2024-11-26 19:27:03.152759] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.749 [2024-11-26 19:27:03.152764] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.749 [2024-11-26 19:27:03.152769] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.749 [2024-11-26 19:27:03.152779] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.749 qpair failed and we were unable to recover it. 00:30:50.749 [2024-11-26 19:27:03.162834] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.749 [2024-11-26 19:27:03.162883] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.749 [2024-11-26 19:27:03.162893] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.749 [2024-11-26 19:27:03.162898] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.749 [2024-11-26 19:27:03.162903] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.749 [2024-11-26 19:27:03.162914] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.749 qpair failed and we were unable to recover it. 00:30:50.749 [2024-11-26 19:27:03.172839] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.749 [2024-11-26 19:27:03.172916] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.749 [2024-11-26 19:27:03.172929] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.749 [2024-11-26 19:27:03.172934] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.749 [2024-11-26 19:27:03.172939] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.172949] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.182832] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.182883] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.182893] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.182899] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.182903] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.182914] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.192733] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.192781] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.192791] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.192797] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.192801] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.192812] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.202894] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.202974] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.202985] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.202990] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.202995] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.203006] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.212895] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.212945] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.212955] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.212963] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.212968] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.212978] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.222996] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.223061] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.223071] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.223077] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.223082] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.223092] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.232961] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.233008] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.233017] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.233023] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.233028] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.233038] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.242985] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.243037] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.243047] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.243052] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.243057] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.243067] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.252991] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.253044] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.253054] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.750 [2024-11-26 19:27:03.253059] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.750 [2024-11-26 19:27:03.253064] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.750 [2024-11-26 19:27:03.253074] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.750 qpair failed and we were unable to recover it. 00:30:50.750 [2024-11-26 19:27:03.263049] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.750 [2024-11-26 19:27:03.263128] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.750 [2024-11-26 19:27:03.263139] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.263144] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.263149] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.263159] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.273080] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.273126] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.273136] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.273141] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.273146] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.273156] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.283104] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.283158] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.283168] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.283173] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.283178] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.283188] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.293012] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.293063] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.293073] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.293078] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.293083] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.293094] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.303030] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.303081] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.303091] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.303096] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.303101] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.303111] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.313184] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.313233] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.313242] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.313248] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.313252] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.313262] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.323238] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.323284] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.323293] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.323299] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.323304] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.323313] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.333218] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.333267] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.333277] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.333282] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.333287] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.333297] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.343283] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.343334] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.343344] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.343351] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.751 [2024-11-26 19:27:03.343356] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.751 [2024-11-26 19:27:03.343366] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.751 qpair failed and we were unable to recover it. 00:30:50.751 [2024-11-26 19:27:03.353291] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.751 [2024-11-26 19:27:03.353339] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.751 [2024-11-26 19:27:03.353348] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.751 [2024-11-26 19:27:03.353354] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.752 [2024-11-26 19:27:03.353358] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.752 [2024-11-26 19:27:03.353368] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.752 qpair failed and we were unable to recover it. 00:30:50.752 [2024-11-26 19:27:03.363235] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:50.752 [2024-11-26 19:27:03.363281] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:50.752 [2024-11-26 19:27:03.363291] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:50.752 [2024-11-26 19:27:03.363296] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:50.752 [2024-11-26 19:27:03.363300] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:50.752 [2024-11-26 19:27:03.363310] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:50.752 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.373332] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.373381] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.373390] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.013 [2024-11-26 19:27:03.373396] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.013 [2024-11-26 19:27:03.373400] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.013 [2024-11-26 19:27:03.373410] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.013 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.383398] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.383450] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.383461] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.013 [2024-11-26 19:27:03.383466] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.013 [2024-11-26 19:27:03.383471] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.013 [2024-11-26 19:27:03.383485] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.013 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.393420] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.393473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.393483] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.013 [2024-11-26 19:27:03.393488] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.013 [2024-11-26 19:27:03.393493] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.013 [2024-11-26 19:27:03.393503] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.013 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.403448] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.403535] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.403545] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.013 [2024-11-26 19:27:03.403551] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.013 [2024-11-26 19:27:03.403557] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.013 [2024-11-26 19:27:03.403567] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.013 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.413487] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.413588] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.413598] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.013 [2024-11-26 19:27:03.413604] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.013 [2024-11-26 19:27:03.413608] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.013 [2024-11-26 19:27:03.413619] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.013 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.423565] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.423627] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.423638] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.013 [2024-11-26 19:27:03.423643] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.013 [2024-11-26 19:27:03.423648] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.013 [2024-11-26 19:27:03.423659] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.013 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.433404] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.433459] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.433469] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.013 [2024-11-26 19:27:03.433475] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.013 [2024-11-26 19:27:03.433480] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.013 [2024-11-26 19:27:03.433490] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.013 qpair failed and we were unable to recover it. 00:30:51.013 [2024-11-26 19:27:03.443560] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.013 [2024-11-26 19:27:03.443604] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.013 [2024-11-26 19:27:03.443613] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.443619] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.443624] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.443634] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.453590] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.453644] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.453662] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.453669] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.453674] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.453688] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.463619] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.463672] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.463684] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.463689] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.463695] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.463706] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.473657] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.473710] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.473724] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.473729] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.473735] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.473747] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.483673] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.483717] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.483728] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.483733] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.483738] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.483749] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.493705] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.493758] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.493768] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.493773] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.493778] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.493789] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.503724] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.503782] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.503792] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.503797] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.503802] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.503812] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.513772] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.513819] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.513828] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.513833] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.513841] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.513851] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.523712] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.523783] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.523793] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.523798] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.523803] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.523813] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.533700] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.533772] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.533781] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.533786] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.533791] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.533801] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.543858] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.543913] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.543923] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.543927] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.543932] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.543942] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.553860] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.553918] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.553928] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.553933] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.553937] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.553948] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.563903] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.563948] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.563958] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.563963] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.014 [2024-11-26 19:27:03.563968] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.014 [2024-11-26 19:27:03.563978] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.014 qpair failed and we were unable to recover it. 00:30:51.014 [2024-11-26 19:27:03.573928] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.014 [2024-11-26 19:27:03.573977] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.014 [2024-11-26 19:27:03.573987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.014 [2024-11-26 19:27:03.573992] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.015 [2024-11-26 19:27:03.573997] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.015 [2024-11-26 19:27:03.574007] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.015 qpair failed and we were unable to recover it. 00:30:51.015 [2024-11-26 19:27:03.583980] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.015 [2024-11-26 19:27:03.584032] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.015 [2024-11-26 19:27:03.584042] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.015 [2024-11-26 19:27:03.584047] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.015 [2024-11-26 19:27:03.584052] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.015 [2024-11-26 19:27:03.584063] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.015 qpair failed and we were unable to recover it. 00:30:51.015 [2024-11-26 19:27:03.593958] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.015 [2024-11-26 19:27:03.594008] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.015 [2024-11-26 19:27:03.594017] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.015 [2024-11-26 19:27:03.594023] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.015 [2024-11-26 19:27:03.594027] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.015 [2024-11-26 19:27:03.594038] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.015 qpair failed and we were unable to recover it. 00:30:51.015 [2024-11-26 19:27:03.604001] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.015 [2024-11-26 19:27:03.604047] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.015 [2024-11-26 19:27:03.604060] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.015 [2024-11-26 19:27:03.604065] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.015 [2024-11-26 19:27:03.604070] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.015 [2024-11-26 19:27:03.604081] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.015 qpair failed and we were unable to recover it. 00:30:51.015 [2024-11-26 19:27:03.614052] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.015 [2024-11-26 19:27:03.614106] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.015 [2024-11-26 19:27:03.614115] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.015 [2024-11-26 19:27:03.614121] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.015 [2024-11-26 19:27:03.614125] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.015 [2024-11-26 19:27:03.614136] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.015 qpair failed and we were unable to recover it. 00:30:51.015 [2024-11-26 19:27:03.624097] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.015 [2024-11-26 19:27:03.624152] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.015 [2024-11-26 19:27:03.624162] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.015 [2024-11-26 19:27:03.624167] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.015 [2024-11-26 19:27:03.624172] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.015 [2024-11-26 19:27:03.624182] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.015 qpair failed and we were unable to recover it. 00:30:51.015 [2024-11-26 19:27:03.633993] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.015 [2024-11-26 19:27:03.634043] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.015 [2024-11-26 19:27:03.634053] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.015 [2024-11-26 19:27:03.634059] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.015 [2024-11-26 19:27:03.634063] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.015 [2024-11-26 19:27:03.634074] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.015 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.644004] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.644050] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.644060] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.644065] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.644072] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.644083] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.654145] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.654211] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.654221] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.654226] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.654230] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.654241] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.664081] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.664144] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.664154] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.664159] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.664164] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.664175] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.674207] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.674253] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.674263] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.674268] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.674274] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.674284] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.684236] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.684327] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.684337] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.684343] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.684347] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.684358] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.694242] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.694293] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.694303] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.694308] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.694313] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.694323] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.704341] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.704392] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.704403] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.704408] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.704413] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.704423] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.714326] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.714375] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.714385] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.714390] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.714395] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.714406] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.724220] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.724269] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.724279] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.724284] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.724289] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.724299] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.734388] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.734439] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.734451] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.734456] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.734461] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.734471] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.744424] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.744472] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.744482] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.744487] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.744492] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.744502] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.754392] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.754473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.754483] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.754488] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.754493] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.754503] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.764428] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.764471] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.764481] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.764487] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.764491] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.764501] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.774499] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.774551] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.774561] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.774572] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.774577] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.774587] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.784538] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.276 [2024-11-26 19:27:03.784589] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.276 [2024-11-26 19:27:03.784599] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.276 [2024-11-26 19:27:03.784604] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.276 [2024-11-26 19:27:03.784609] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.276 [2024-11-26 19:27:03.784619] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.276 qpair failed and we were unable to recover it. 00:30:51.276 [2024-11-26 19:27:03.794402] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.794452] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.794461] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.794467] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.794472] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.794482] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.804537] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.804586] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.804595] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.804600] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.804605] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.804615] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.814607] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.814658] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.814669] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.814674] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.814679] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.814692] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.824629] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.824677] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.824687] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.824692] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.824696] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.824707] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.834674] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.834723] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.834733] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.834738] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.834743] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.834753] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.844645] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.844691] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.844701] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.844707] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.844711] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.844722] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.854723] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.854774] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.854784] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.854789] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.854794] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.854805] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.864741] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.864799] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.864809] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.864814] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.864819] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.864830] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.874758] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.874807] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.874817] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.874822] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.874827] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.874838] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.884743] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.884788] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.884798] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.884803] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.884808] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.884819] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.277 [2024-11-26 19:27:03.894875] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.277 [2024-11-26 19:27:03.894926] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.277 [2024-11-26 19:27:03.894936] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.277 [2024-11-26 19:27:03.894941] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.277 [2024-11-26 19:27:03.894946] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.277 [2024-11-26 19:27:03.894957] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.277 qpair failed and we were unable to recover it. 00:30:51.538 [2024-11-26 19:27:03.904845] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.538 [2024-11-26 19:27:03.904899] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.538 [2024-11-26 19:27:03.904909] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.538 [2024-11-26 19:27:03.904917] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.538 [2024-11-26 19:27:03.904922] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.538 [2024-11-26 19:27:03.904932] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.538 qpair failed and we were unable to recover it. 00:30:51.538 [2024-11-26 19:27:03.914727] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.538 [2024-11-26 19:27:03.914773] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.538 [2024-11-26 19:27:03.914783] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.538 [2024-11-26 19:27:03.914789] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.538 [2024-11-26 19:27:03.914793] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.538 [2024-11-26 19:27:03.914804] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.538 qpair failed and we were unable to recover it. 00:30:51.538 [2024-11-26 19:27:03.924838] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.538 [2024-11-26 19:27:03.924877] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.538 [2024-11-26 19:27:03.924888] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.538 [2024-11-26 19:27:03.924893] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.538 [2024-11-26 19:27:03.924898] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.538 [2024-11-26 19:27:03.924909] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.538 qpair failed and we were unable to recover it. 00:30:51.538 [2024-11-26 19:27:03.934952] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.538 [2024-11-26 19:27:03.935006] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.538 [2024-11-26 19:27:03.935016] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.538 [2024-11-26 19:27:03.935021] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.538 [2024-11-26 19:27:03.935026] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.538 [2024-11-26 19:27:03.935037] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.538 qpair failed and we were unable to recover it. 00:30:51.538 [2024-11-26 19:27:03.944950] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.538 [2024-11-26 19:27:03.945048] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.538 [2024-11-26 19:27:03.945059] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.538 [2024-11-26 19:27:03.945064] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.538 [2024-11-26 19:27:03.945069] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.538 [2024-11-26 19:27:03.945082] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.538 qpair failed and we were unable to recover it. 00:30:51.538 [2024-11-26 19:27:03.954967] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.538 [2024-11-26 19:27:03.955015] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.538 [2024-11-26 19:27:03.955025] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.538 [2024-11-26 19:27:03.955030] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.538 [2024-11-26 19:27:03.955034] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.538 [2024-11-26 19:27:03.955045] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.538 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:03.964958] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:03.965037] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:03.965047] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:03.965052] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:03.965056] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:03.965067] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:03.975022] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:03.975070] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:03.975080] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:03.975085] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:03.975090] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:03.975100] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:03.985075] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:03.985128] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:03.985138] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:03.985143] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:03.985148] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:03.985158] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:03.995043] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:03.995092] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:03.995102] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:03.995107] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:03.995112] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:03.995122] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.005077] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.005127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.005136] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.005142] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.005147] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.005157] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.015013] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.015113] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.015124] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.015130] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.015135] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.015147] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.025190] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.025244] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.025254] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.025259] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.025264] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.025275] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.035059] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.035107] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.035120] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.035125] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.035130] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.035140] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.045178] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.045233] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.045244] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.045249] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.045254] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.045265] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.055268] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.055320] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.055330] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.055335] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.055340] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.055351] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.065274] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.065326] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.065336] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.065341] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.065346] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.065356] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.075314] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.075399] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.075409] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.075414] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.075422] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.075433] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.085139] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.539 [2024-11-26 19:27:04.085182] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.539 [2024-11-26 19:27:04.085192] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.539 [2024-11-26 19:27:04.085198] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.539 [2024-11-26 19:27:04.085203] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.539 [2024-11-26 19:27:04.085214] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.539 qpair failed and we were unable to recover it. 00:30:51.539 [2024-11-26 19:27:04.095350] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.540 [2024-11-26 19:27:04.095400] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.540 [2024-11-26 19:27:04.095410] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.540 [2024-11-26 19:27:04.095415] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.540 [2024-11-26 19:27:04.095420] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.540 [2024-11-26 19:27:04.095431] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.540 qpair failed and we were unable to recover it. 00:30:51.540 [2024-11-26 19:27:04.105372] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.540 [2024-11-26 19:27:04.105420] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.540 [2024-11-26 19:27:04.105429] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.540 [2024-11-26 19:27:04.105435] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.540 [2024-11-26 19:27:04.105439] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.540 [2024-11-26 19:27:04.105450] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.540 qpair failed and we were unable to recover it. 00:30:51.540 [2024-11-26 19:27:04.115374] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.540 [2024-11-26 19:27:04.115435] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.540 [2024-11-26 19:27:04.115444] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.540 [2024-11-26 19:27:04.115450] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.540 [2024-11-26 19:27:04.115454] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.540 [2024-11-26 19:27:04.115465] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.540 qpair failed and we were unable to recover it. 00:30:51.540 [2024-11-26 19:27:04.125382] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.540 [2024-11-26 19:27:04.125430] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.540 [2024-11-26 19:27:04.125440] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.540 [2024-11-26 19:27:04.125446] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.540 [2024-11-26 19:27:04.125450] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.540 [2024-11-26 19:27:04.125460] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.540 qpair failed and we were unable to recover it. 00:30:51.540 [2024-11-26 19:27:04.135508] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.540 [2024-11-26 19:27:04.135561] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.540 [2024-11-26 19:27:04.135571] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.540 [2024-11-26 19:27:04.135576] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.540 [2024-11-26 19:27:04.135580] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.540 [2024-11-26 19:27:04.135590] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.540 qpair failed and we were unable to recover it. 00:30:51.540 [2024-11-26 19:27:04.145497] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.540 [2024-11-26 19:27:04.145549] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.540 [2024-11-26 19:27:04.145558] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.540 [2024-11-26 19:27:04.145564] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.540 [2024-11-26 19:27:04.145568] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.540 [2024-11-26 19:27:04.145578] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.540 qpair failed and we were unable to recover it. 00:30:51.540 [2024-11-26 19:27:04.155501] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.540 [2024-11-26 19:27:04.155554] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.540 [2024-11-26 19:27:04.155564] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.540 [2024-11-26 19:27:04.155569] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.540 [2024-11-26 19:27:04.155574] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.540 [2024-11-26 19:27:04.155584] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.540 qpair failed and we were unable to recover it. 00:30:51.801 [2024-11-26 19:27:04.165447] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.801 [2024-11-26 19:27:04.165491] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.801 [2024-11-26 19:27:04.165504] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.801 [2024-11-26 19:27:04.165509] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.165514] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.165525] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.175574] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.175627] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.175637] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.175643] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.175647] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.175658] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.185596] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.185650] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.185660] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.185665] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.185670] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.185681] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.195596] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.195641] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.195651] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.195657] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.195661] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.195672] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.205464] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.205540] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.205550] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.205555] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.205562] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.205573] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.215591] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.215683] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.215692] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.215698] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.215704] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.215714] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.225683] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.225743] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.225752] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.225758] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.225763] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.225773] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.235699] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.235750] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.235759] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.235764] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.235769] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.235780] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.245732] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.245774] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.245784] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.245789] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.245794] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.245805] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.255784] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.255836] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.255846] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.255851] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.255856] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.255877] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.265834] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.265889] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.265899] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.265905] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.265909] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.265920] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.275800] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.275844] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.275854] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.275859] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.275868] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.275879] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.285806] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.285848] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.285857] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.802 [2024-11-26 19:27:04.285867] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.802 [2024-11-26 19:27:04.285872] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.802 [2024-11-26 19:27:04.285882] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.802 qpair failed and we were unable to recover it. 00:30:51.802 [2024-11-26 19:27:04.295900] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.802 [2024-11-26 19:27:04.295988] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.802 [2024-11-26 19:27:04.296000] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.296006] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.296010] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.296021] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.305930] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.305977] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.305987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.305992] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.305998] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.306008] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.315957] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.316003] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.316012] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.316018] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.316023] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.316033] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.325940] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.325985] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.325995] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.326001] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.326005] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.326016] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.336016] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.336114] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.336126] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.336134] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.336139] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.336150] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.346011] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.346059] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.346069] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.346074] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.346079] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.346089] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.355985] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.356034] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.356045] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.356051] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.356056] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.356067] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.365923] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.365972] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.365982] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.365987] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.365992] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.366003] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.376114] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.376198] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.376208] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.376213] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.376218] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.376231] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.386026] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.386073] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.386083] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.386088] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.386093] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.386104] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.396048] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.396102] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.396112] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.396117] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.396121] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.396132] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.406049] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.406132] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.406141] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.406147] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.406151] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.406162] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:51.803 [2024-11-26 19:27:04.416282] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:51.803 [2024-11-26 19:27:04.416334] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:51.803 [2024-11-26 19:27:04.416343] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:51.803 [2024-11-26 19:27:04.416348] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:51.803 [2024-11-26 19:27:04.416353] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:51.803 [2024-11-26 19:27:04.416363] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:51.803 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.426263] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.426325] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.426334] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.426340] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.426345] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.426355] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.436277] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.436330] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.436346] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.436352] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.436357] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.436371] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.446261] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.446312] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.446323] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.446328] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.446333] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.446343] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.456330] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.456381] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.456391] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.456396] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.456401] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.456411] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.466354] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.466405] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.466415] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.466423] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.466428] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.466438] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.476434] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.476506] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.476516] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.476521] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.476526] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.476537] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.486371] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.486417] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.486428] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.486433] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.486438] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.486448] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.496454] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.496509] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.496519] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.496524] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.496530] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.496540] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.506476] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.506531] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.506540] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.506546] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.064 [2024-11-26 19:27:04.506551] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.064 [2024-11-26 19:27:04.506567] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.064 qpair failed and we were unable to recover it. 00:30:52.064 [2024-11-26 19:27:04.516458] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.064 [2024-11-26 19:27:04.516510] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.064 [2024-11-26 19:27:04.516520] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.064 [2024-11-26 19:27:04.516525] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.516530] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.516540] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.526499] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.526562] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.526572] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.526577] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.526582] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.526592] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.536528] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.536579] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.536589] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.536595] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.536599] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.536610] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.546567] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.546620] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.546630] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.546635] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.546640] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.546650] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.556602] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.556655] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.556665] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.556671] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.556675] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.556685] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.566577] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.566637] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.566647] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.566652] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.566657] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.566667] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.576645] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.576695] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.576706] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.576711] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.576716] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.576726] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.586574] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.586626] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.586637] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.586643] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.586648] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.586658] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.596701] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.596748] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.596762] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.596767] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.596772] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.596782] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.606700] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.606781] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.606791] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.606797] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.606801] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.606813] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.616771] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.616822] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.616833] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.616838] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.616843] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.616853] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.626813] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.626906] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.626917] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.626922] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.626928] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.626939] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.636823] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.636874] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.636883] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.636889] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.636896] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.065 [2024-11-26 19:27:04.636907] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.065 qpair failed and we were unable to recover it. 00:30:52.065 [2024-11-26 19:27:04.646788] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.065 [2024-11-26 19:27:04.646827] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.065 [2024-11-26 19:27:04.646837] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.065 [2024-11-26 19:27:04.646842] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.065 [2024-11-26 19:27:04.646847] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.066 [2024-11-26 19:27:04.646857] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.066 qpair failed and we were unable to recover it. 00:30:52.066 [2024-11-26 19:27:04.656886] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.066 [2024-11-26 19:27:04.656937] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.066 [2024-11-26 19:27:04.656948] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.066 [2024-11-26 19:27:04.656954] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.066 [2024-11-26 19:27:04.656959] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.066 [2024-11-26 19:27:04.656970] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.066 qpair failed and we were unable to recover it. 00:30:52.066 [2024-11-26 19:27:04.666860] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.066 [2024-11-26 19:27:04.666914] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.066 [2024-11-26 19:27:04.666924] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.066 [2024-11-26 19:27:04.666929] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.066 [2024-11-26 19:27:04.666934] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.066 [2024-11-26 19:27:04.666945] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.066 qpair failed and we were unable to recover it. 00:30:52.066 [2024-11-26 19:27:04.676954] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.066 [2024-11-26 19:27:04.676999] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.066 [2024-11-26 19:27:04.677009] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.066 [2024-11-26 19:27:04.677014] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.066 [2024-11-26 19:27:04.677019] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.066 [2024-11-26 19:27:04.677029] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.066 qpair failed and we were unable to recover it. 00:30:52.327 [2024-11-26 19:27:04.686927] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.327 [2024-11-26 19:27:04.686981] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.327 [2024-11-26 19:27:04.686991] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.327 [2024-11-26 19:27:04.686997] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.327 [2024-11-26 19:27:04.687001] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.327 [2024-11-26 19:27:04.687012] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.327 qpair failed and we were unable to recover it. 00:30:52.327 [2024-11-26 19:27:04.696966] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.327 [2024-11-26 19:27:04.697030] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.327 [2024-11-26 19:27:04.697039] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.697044] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.697049] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.697060] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.706946] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.706996] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.707006] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.707011] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.707016] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.707026] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.717065] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.717158] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.717168] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.717173] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.717178] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.717188] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.727008] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.727056] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.727068] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.727073] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.727078] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.727089] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.736976] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.737027] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.737037] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.737042] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.737047] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.737057] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.747150] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.747196] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.747206] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.747212] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.747216] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.747226] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.757151] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.757199] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.757209] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.757214] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.757219] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.757230] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.767135] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.767175] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.767185] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.767190] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.767197] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.767207] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.777108] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.777207] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.777218] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.777223] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.777228] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.777238] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.787294] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.787386] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.787396] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.787401] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.787406] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.787416] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.797244] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.797302] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.797312] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.797317] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.797322] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.797332] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.807239] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.807297] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.807307] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.807313] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.807317] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.807327] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.817195] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.817246] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.817257] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.817263] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.328 [2024-11-26 19:27:04.817267] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.328 [2024-11-26 19:27:04.817278] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.328 qpair failed and we were unable to recover it. 00:30:52.328 [2024-11-26 19:27:04.827215] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.328 [2024-11-26 19:27:04.827268] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.328 [2024-11-26 19:27:04.827279] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.328 [2024-11-26 19:27:04.827284] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.827289] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.827300] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.837392] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.837445] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.837455] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.837461] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.837465] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.837475] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.847321] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.847371] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.847381] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.847386] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.847391] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.847401] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.857348] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.857397] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.857410] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.857415] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.857420] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.857430] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.867473] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.867530] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.867540] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.867546] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.867551] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.867561] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.877355] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.877404] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.877413] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.877418] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.877423] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.877434] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.887475] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.887525] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.887534] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.887540] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.887544] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.887554] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.897563] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.897612] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.897622] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.897630] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.897635] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.897645] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.907446] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.907499] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.907509] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.907514] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.907519] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.907530] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.917604] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.917658] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.917668] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.917673] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.917678] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.917688] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.927600] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.927654] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.927663] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.927669] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.927673] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.927684] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.937685] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.937738] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.937747] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.937752] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.937757] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.937770] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.329 [2024-11-26 19:27:04.947655] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.329 [2024-11-26 19:27:04.947704] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.329 [2024-11-26 19:27:04.947714] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.329 [2024-11-26 19:27:04.947720] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.329 [2024-11-26 19:27:04.947724] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.329 [2024-11-26 19:27:04.947734] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.329 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:04.957706] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:04.957753] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:04.957763] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:04.957769] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:04.957774] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:04.957784] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:04.967699] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:04.967742] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:04.967752] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:04.967758] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:04.967763] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:04.967773] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:04.977763] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:04.977813] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:04.977823] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:04.977828] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:04.977833] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:04.977845] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:04.987776] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:04.987832] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:04.987842] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:04.987847] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:04.987852] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:04.987867] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:04.997711] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:04.997759] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:04.997770] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:04.997776] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:04.997780] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:04.997791] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:05.007657] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:05.007700] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:05.007712] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:05.007717] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:05.007722] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:05.007733] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:05.017874] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:05.017961] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:05.017971] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:05.017977] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:05.017982] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:05.017993] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:05.027910] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:05.027963] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:05.027974] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:05.027983] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:05.027988] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:05.028000] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:05.037778] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:05.037829] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:05.037838] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:05.037844] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:05.037848] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:05.037858] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:05.047889] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:05.047982] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:05.047993] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:05.047999] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:05.048003] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:05.048014] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:05.057951] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:05.058005] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:05.058014] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:05.058020] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:05.058024] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:05.058035] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.591 [2024-11-26 19:27:05.067970] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.591 [2024-11-26 19:27:05.068022] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.591 [2024-11-26 19:27:05.068032] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.591 [2024-11-26 19:27:05.068037] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.591 [2024-11-26 19:27:05.068042] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.591 [2024-11-26 19:27:05.068055] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.591 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.078017] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.078063] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.078073] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.078078] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.078083] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.078093] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.088011] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.088049] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.088058] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.088064] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.088068] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.088079] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.098087] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.098149] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.098159] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.098164] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.098169] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.098179] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.108091] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.108140] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.108149] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.108155] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.108160] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.108170] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.118129] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.118182] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.118192] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.118197] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.118202] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.118213] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.127998] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.128038] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.128048] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.128054] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.128058] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.128069] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.138186] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.138236] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.138246] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.138251] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.138256] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.138266] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.148097] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.148152] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.148161] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.148167] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.148171] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.148182] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.158292] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.158355] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.158367] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.158373] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.158378] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.158388] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.168227] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.168310] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.168319] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.168325] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.168330] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.168341] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.178285] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.178337] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.178346] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.178352] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.178356] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.178367] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.188323] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.188373] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.188383] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.188388] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.188393] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.188403] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.198336] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.592 [2024-11-26 19:27:05.198387] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.592 [2024-11-26 19:27:05.198397] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.592 [2024-11-26 19:27:05.198402] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.592 [2024-11-26 19:27:05.198413] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.592 [2024-11-26 19:27:05.198423] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.592 qpair failed and we were unable to recover it. 00:30:52.592 [2024-11-26 19:27:05.208304] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.593 [2024-11-26 19:27:05.208345] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.593 [2024-11-26 19:27:05.208355] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.593 [2024-11-26 19:27:05.208360] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.593 [2024-11-26 19:27:05.208365] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.593 [2024-11-26 19:27:05.208376] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.593 qpair failed and we were unable to recover it. 00:30:52.853 [2024-11-26 19:27:05.218401] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.853 [2024-11-26 19:27:05.218452] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.853 [2024-11-26 19:27:05.218462] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.853 [2024-11-26 19:27:05.218467] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.853 [2024-11-26 19:27:05.218472] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.853 [2024-11-26 19:27:05.218482] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.853 qpair failed and we were unable to recover it. 00:30:52.853 [2024-11-26 19:27:05.228437] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.853 [2024-11-26 19:27:05.228537] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.853 [2024-11-26 19:27:05.228548] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.853 [2024-11-26 19:27:05.228553] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.853 [2024-11-26 19:27:05.228558] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.853 [2024-11-26 19:27:05.228568] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.853 qpair failed and we were unable to recover it. 00:30:52.853 [2024-11-26 19:27:05.238436] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.853 [2024-11-26 19:27:05.238517] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.853 [2024-11-26 19:27:05.238527] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.853 [2024-11-26 19:27:05.238532] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.853 [2024-11-26 19:27:05.238536] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.853 [2024-11-26 19:27:05.238547] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.853 qpair failed and we were unable to recover it. 00:30:52.853 [2024-11-26 19:27:05.248422] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.853 [2024-11-26 19:27:05.248493] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.853 [2024-11-26 19:27:05.248502] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.853 [2024-11-26 19:27:05.248508] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.853 [2024-11-26 19:27:05.248513] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.853 [2024-11-26 19:27:05.248523] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.853 qpair failed and we were unable to recover it. 00:30:52.853 [2024-11-26 19:27:05.258502] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.853 [2024-11-26 19:27:05.258555] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.853 [2024-11-26 19:27:05.258564] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.258569] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.258574] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.258584] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.268541] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.268596] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.268605] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.268610] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.268615] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.268625] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.278552] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.278604] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.278614] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.278619] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.278623] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.278633] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.288542] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.288603] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.288615] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.288620] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.288625] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.288635] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.298492] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.298548] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.298558] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.298563] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.298567] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.298577] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.308636] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.308717] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.308726] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.308731] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.308736] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.308746] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.318626] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.318678] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.318697] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.318703] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.318708] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.318722] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.328513] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.328559] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.328577] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.328584] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.328593] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.328607] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.338717] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.338817] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.338836] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.338842] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.338847] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.338867] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.348764] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.348814] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.854 [2024-11-26 19:27:05.348826] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.854 [2024-11-26 19:27:05.348831] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.854 [2024-11-26 19:27:05.348836] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.854 [2024-11-26 19:27:05.348847] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.854 qpair failed and we were unable to recover it. 00:30:52.854 [2024-11-26 19:27:05.358602] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.854 [2024-11-26 19:27:05.358652] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.358664] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.358670] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.358674] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.358685] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.368739] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.368778] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.368789] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.368794] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.368799] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.368809] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.378824] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.378876] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.378887] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.378892] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.378897] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.378907] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.388877] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.388929] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.388938] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.388944] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.388948] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.388959] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.398705] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.398750] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.398760] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.398765] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.398770] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.398781] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.408849] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.408937] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.408947] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.408953] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.408958] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.408968] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.418934] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.418982] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.418994] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.419000] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.419004] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.419015] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.429012] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.429086] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.429096] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.429101] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.429106] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.429116] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.438810] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.438849] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.438859] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.438868] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.438873] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.438884] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.448973] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.449016] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.449026] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.449031] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.449036] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.449046] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.459119] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.855 [2024-11-26 19:27:05.459172] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.855 [2024-11-26 19:27:05.459181] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.855 [2024-11-26 19:27:05.459189] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.855 [2024-11-26 19:27:05.459194] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.855 [2024-11-26 19:27:05.459204] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.855 qpair failed and we were unable to recover it. 00:30:52.855 [2024-11-26 19:27:05.469130] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:52.856 [2024-11-26 19:27:05.469180] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:52.856 [2024-11-26 19:27:05.469190] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:52.856 [2024-11-26 19:27:05.469195] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:52.856 [2024-11-26 19:27:05.469200] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:52.856 [2024-11-26 19:27:05.469210] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:52.856 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.479053] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.479096] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.479105] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.479110] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.479115] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.479125] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.489086] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.489132] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.489141] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.489146] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.489151] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.489161] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.499165] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.499213] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.499223] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.499228] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.499233] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.499245] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.509065] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.509114] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.509124] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.509129] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.509134] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.509145] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.519163] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.519206] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.519217] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.519222] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.519226] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.519237] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.529203] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.529248] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.529258] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.529263] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.529268] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.529278] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.539171] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.539237] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.539246] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.539252] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.539256] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.539266] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.549351] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.549406] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.549416] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.549421] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.549425] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.549436] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.559301] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.559345] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.559355] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.559360] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.559364] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.559375] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.569262] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.569304] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.569314] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.569319] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.569324] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.569334] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.579400] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.579451] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.579461] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.579466] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.579472] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.579482] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.589408] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.589495] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.116 [2024-11-26 19:27:05.589505] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.116 [2024-11-26 19:27:05.589513] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.116 [2024-11-26 19:27:05.589518] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.116 [2024-11-26 19:27:05.589528] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.116 qpair failed and we were unable to recover it. 00:30:53.116 [2024-11-26 19:27:05.599368] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.116 [2024-11-26 19:27:05.599413] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.599423] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.599428] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.599433] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.599444] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.609443] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.609528] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.609538] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.609544] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.609549] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.609559] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.619541] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.619599] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.619609] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.619614] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.619618] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.619628] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.629520] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.629571] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.629580] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.629586] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.629590] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.629603] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.639383] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.639440] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.639450] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.639455] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.639459] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.639469] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.649485] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.649533] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.649543] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.649548] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.649553] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.649563] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.659618] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.659714] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.659725] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.659730] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.659735] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.659745] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.669684] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.669734] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.669743] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.669748] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.669753] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.669763] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.679610] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.679657] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.679667] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.679672] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.679677] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.679687] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.689493] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.689537] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.689548] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.689553] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.689558] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.689568] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.699684] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.699740] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.699751] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.699756] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.699761] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.699771] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.709612] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.709671] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.709681] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.709686] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.709691] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.709702] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.719713] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.117 [2024-11-26 19:27:05.719802] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.117 [2024-11-26 19:27:05.719815] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.117 [2024-11-26 19:27:05.719821] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.117 [2024-11-26 19:27:05.719826] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.117 [2024-11-26 19:27:05.719837] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.117 qpair failed and we were unable to recover it. 00:30:53.117 [2024-11-26 19:27:05.729765] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.118 [2024-11-26 19:27:05.729805] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.118 [2024-11-26 19:27:05.729815] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.118 [2024-11-26 19:27:05.729821] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.118 [2024-11-26 19:27:05.729826] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.118 [2024-11-26 19:27:05.729836] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.118 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.739821] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.739876] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.739886] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.739891] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.739896] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.739907] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.749853] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.749954] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.749965] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.749971] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.749976] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.749986] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.759812] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.759855] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.759869] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.759875] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.759882] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.759893] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.769876] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.769925] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.769935] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.769941] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.769945] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.769956] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.779932] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.779981] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.779991] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.779996] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.780001] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.780011] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.789954] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.790003] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.790013] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.790018] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.790022] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.790033] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.799955] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.799997] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.800006] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.800012] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.800017] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.800027] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.809960] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.810038] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.810048] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.810053] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.810058] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.810069] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.820040] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.820092] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.820101] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.820106] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.820111] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.820122] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.830051] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.830103] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.830112] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.830118] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.830122] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.830132] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.840027] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.840066] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.840076] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.840082] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.840087] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.840098] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.849950] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.849994] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.850006] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.850012] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.850016] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.850026] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.860158] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.860208] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.860218] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.860223] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.860228] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.860238] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.870197] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.378 [2024-11-26 19:27:05.870249] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.378 [2024-11-26 19:27:05.870258] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.378 [2024-11-26 19:27:05.870263] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.378 [2024-11-26 19:27:05.870268] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.378 [2024-11-26 19:27:05.870279] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.378 qpair failed and we were unable to recover it. 00:30:53.378 [2024-11-26 19:27:05.880175] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.880216] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.880225] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.880230] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.880235] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.379 [2024-11-26 19:27:05.880245] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.890177] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.890224] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.890234] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.890239] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.890249] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.379 [2024-11-26 19:27:05.890260] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.900271] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.900320] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.900331] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.900338] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.900344] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.379 [2024-11-26 19:27:05.900355] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.910183] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.910233] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.910243] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.910248] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.910253] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.379 [2024-11-26 19:27:05.910264] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.920272] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.920314] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.920323] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.920328] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.920333] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.379 [2024-11-26 19:27:05.920343] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.930303] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.930390] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.930400] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.930406] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.930411] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb788000b90 00:30:53.379 [2024-11-26 19:27:05.930422] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.940456] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.940572] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.940637] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.940664] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.940685] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb784000b90 00:30:53.379 [2024-11-26 19:27:05.940738] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.950435] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.950512] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.950544] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.950560] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.950575] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb784000b90 00:30:53.379 [2024-11-26 19:27:05.950608] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.960405] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.960511] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.960576] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.960602] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.960623] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb790000b90 00:30:53.379 [2024-11-26 19:27:05.960679] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.970401] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.970468] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.970496] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.970512] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.970525] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fb790000b90 00:30:53.379 [2024-11-26 19:27:05.970556] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.980485] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.980550] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.980575] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.980584] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.980592] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1a26490 00:30:53.379 [2024-11-26 19:27:05.980612] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.990521] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:30:53.379 [2024-11-26 19:27:05.990581] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:30:53.379 [2024-11-26 19:27:05.990606] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:30:53.379 [2024-11-26 19:27:05.990615] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:53.379 [2024-11-26 19:27:05.990623] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1a26490 00:30:53.379 [2024-11-26 19:27:05.990643] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:30:53.379 qpair failed and we were unable to recover it. 00:30:53.379 [2024-11-26 19:27:05.990813] nvme_ctrlr.c:4518:nvme_ctrlr_keep_alive: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Submitting Keep Alive failed 00:30:53.379 A controller has encountered a failure and is being reset. 00:30:53.379 [2024-11-26 19:27:05.990950] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1a23030 (9): Bad file descriptor 00:30:53.639 Controller properly reset. 00:30:53.639 Initializing NVMe Controllers 00:30:53.639 Attaching to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:30:53.639 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:30:53.639 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 0 00:30:53.639 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 1 00:30:53.639 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 2 00:30:53.639 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 3 00:30:53.639 Initialization complete. Launching workers. 00:30:53.639 Starting thread on core 1 00:30:53.639 Starting thread on core 2 00:30:53.639 Starting thread on core 3 00:30:53.639 Starting thread on core 0 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@51 -- # sync 00:30:53.639 00:30:53.639 real 0m11.490s 00:30:53.639 user 0m21.331s 00:30:53.639 sys 0m3.904s 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:30:53.639 ************************************ 00:30:53.639 END TEST nvmf_target_disconnect_tc2 00:30:53.639 ************************************ 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@72 -- # '[' -n '' ']' 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@76 -- # trap - SIGINT SIGTERM EXIT 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@77 -- # nvmftestfini 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@516 -- # nvmfcleanup 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@121 -- # sync 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@124 -- # set +e 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:53.639 rmmod nvme_tcp 00:30:53.639 rmmod nvme_fabrics 00:30:53.639 rmmod nvme_keyring 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@128 -- # set -e 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@129 -- # return 0 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@517 -- # '[' -n 4065791 ']' 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@518 -- # killprocess 4065791 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@954 -- # '[' -z 4065791 ']' 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@958 -- # kill -0 4065791 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@959 -- # uname 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4065791 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@960 -- # process_name=reactor_4 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@964 -- # '[' reactor_4 = sudo ']' 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4065791' 00:30:53.639 killing process with pid 4065791 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@973 -- # kill 4065791 00:30:53.639 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@978 -- # wait 4065791 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@297 -- # iptr 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@791 -- # iptables-restore 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@791 -- # iptables-save 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:53.899 19:27:06 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:56.441 19:27:08 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:56.441 00:30:56.441 real 0m22.634s 00:30:56.441 user 0m49.669s 00:30:56.441 sys 0m10.611s 00:30:56.441 19:27:08 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:56.441 19:27:08 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:30:56.441 ************************************ 00:30:56.441 END TEST nvmf_target_disconnect 00:30:56.441 ************************************ 00:30:56.441 19:27:08 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:30:56.441 00:30:56.441 real 6m48.418s 00:30:56.441 user 11m24.885s 00:30:56.441 sys 2m25.849s 00:30:56.441 19:27:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:56.441 19:27:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:30:56.441 ************************************ 00:30:56.441 END TEST nvmf_host 00:30:56.442 ************************************ 00:30:56.442 19:27:08 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ tcp = \t\c\p ]] 00:30:56.442 19:27:08 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ 0 -eq 0 ]] 00:30:56.442 19:27:08 nvmf_tcp -- nvmf/nvmf.sh@20 -- # run_test nvmf_target_core_interrupt_mode /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:30:56.442 19:27:08 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:30:56.442 19:27:08 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:56.442 19:27:08 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:30:56.442 ************************************ 00:30:56.442 START TEST nvmf_target_core_interrupt_mode 00:30:56.442 ************************************ 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:30:56.442 * Looking for test storage... 00:30:56.442 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1693 -- # lcov --version 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # IFS=.-: 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # read -ra ver1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # IFS=.-: 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # read -ra ver2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@338 -- # local 'op=<' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@340 -- # ver1_l=2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@341 -- # ver2_l=1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@344 -- # case "$op" in 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@345 -- # : 1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # decimal 1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # ver1[v]=1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # decimal 2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # ver2[v]=2 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # return 0 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:30:56.442 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.442 --rc genhtml_branch_coverage=1 00:30:56.442 --rc genhtml_function_coverage=1 00:30:56.442 --rc genhtml_legend=1 00:30:56.442 --rc geninfo_all_blocks=1 00:30:56.442 --rc geninfo_unexecuted_blocks=1 00:30:56.442 00:30:56.442 ' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:30:56.442 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.442 --rc genhtml_branch_coverage=1 00:30:56.442 --rc genhtml_function_coverage=1 00:30:56.442 --rc genhtml_legend=1 00:30:56.442 --rc geninfo_all_blocks=1 00:30:56.442 --rc geninfo_unexecuted_blocks=1 00:30:56.442 00:30:56.442 ' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:30:56.442 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.442 --rc genhtml_branch_coverage=1 00:30:56.442 --rc genhtml_function_coverage=1 00:30:56.442 --rc genhtml_legend=1 00:30:56.442 --rc geninfo_all_blocks=1 00:30:56.442 --rc geninfo_unexecuted_blocks=1 00:30:56.442 00:30:56.442 ' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:30:56.442 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.442 --rc genhtml_branch_coverage=1 00:30:56.442 --rc genhtml_function_coverage=1 00:30:56.442 --rc genhtml_legend=1 00:30:56.442 --rc geninfo_all_blocks=1 00:30:56.442 --rc geninfo_unexecuted_blocks=1 00:30:56.442 00:30:56.442 ' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # uname -s 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@15 -- # shopt -s extglob 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@5 -- # export PATH 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@51 -- # : 0 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:56.442 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:30:56.443 ************************************ 00:30:56.443 START TEST nvmf_abort 00:30:56.443 ************************************ 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:30:56.443 * Looking for test storage... 00:30:56.443 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1693 -- # lcov --version 00:30:56.443 19:27:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:56.443 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:30:56.706 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.706 --rc genhtml_branch_coverage=1 00:30:56.706 --rc genhtml_function_coverage=1 00:30:56.706 --rc genhtml_legend=1 00:30:56.706 --rc geninfo_all_blocks=1 00:30:56.706 --rc geninfo_unexecuted_blocks=1 00:30:56.706 00:30:56.706 ' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:30:56.706 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.706 --rc genhtml_branch_coverage=1 00:30:56.706 --rc genhtml_function_coverage=1 00:30:56.706 --rc genhtml_legend=1 00:30:56.706 --rc geninfo_all_blocks=1 00:30:56.706 --rc geninfo_unexecuted_blocks=1 00:30:56.706 00:30:56.706 ' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:30:56.706 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.706 --rc genhtml_branch_coverage=1 00:30:56.706 --rc genhtml_function_coverage=1 00:30:56.706 --rc genhtml_legend=1 00:30:56.706 --rc geninfo_all_blocks=1 00:30:56.706 --rc geninfo_unexecuted_blocks=1 00:30:56.706 00:30:56.706 ' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:30:56.706 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:56.706 --rc genhtml_branch_coverage=1 00:30:56.706 --rc genhtml_function_coverage=1 00:30:56.706 --rc genhtml_legend=1 00:30:56.706 --rc geninfo_all_blocks=1 00:30:56.706 --rc geninfo_unexecuted_blocks=1 00:30:56.706 00:30:56.706 ' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@476 -- # prepare_net_devs 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@438 -- # local -g is_hw=no 00:30:56.706 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # remove_spdk_ns 00:30:56.707 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:56.707 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:56.707 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:56.707 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:30:56.707 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:30:56.707 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:30:56.707 19:27:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:04.856 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:31:04.857 Found 0000:31:00.0 (0x8086 - 0x159b) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:31:04.857 Found 0000:31:00.1 (0x8086 - 0x159b) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:31:04.857 Found net devices under 0000:31:00.0: cvl_0_0 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:31:04.857 Found net devices under 0000:31:00.1: cvl_0_1 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # is_hw=yes 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:04.857 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:05.118 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:05.118 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.687 ms 00:31:05.118 00:31:05.118 --- 10.0.0.2 ping statistics --- 00:31:05.118 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:05.118 rtt min/avg/max/mdev = 0.687/0.687/0.687/0.000 ms 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:05.118 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:05.118 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.334 ms 00:31:05.118 00:31:05.118 --- 10.0.0.1 ping statistics --- 00:31:05.118 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:05.118 rtt min/avg/max/mdev = 0.334/0.334/0.334/0.000 ms 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@450 -- # return 0 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@509 -- # nvmfpid=4072458 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@510 -- # waitforlisten 4072458 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@835 -- # '[' -z 4072458 ']' 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:05.118 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:05.118 19:27:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:05.379 [2024-11-26 19:27:17.765747] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:31:05.379 [2024-11-26 19:27:17.766744] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:31:05.379 [2024-11-26 19:27:17.766783] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:05.379 [2024-11-26 19:27:17.871215] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:31:05.379 [2024-11-26 19:27:17.922040] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:05.379 [2024-11-26 19:27:17.922088] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:05.379 [2024-11-26 19:27:17.922097] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:05.379 [2024-11-26 19:27:17.922104] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:05.379 [2024-11-26 19:27:17.922110] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:05.379 [2024-11-26 19:27:17.923887] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:31:05.379 [2024-11-26 19:27:17.924084] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:31:05.379 [2024-11-26 19:27:17.924178] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:05.379 [2024-11-26 19:27:18.001288] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:31:05.379 [2024-11-26 19:27:18.001322] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:31:05.379 [2024-11-26 19:27:18.002011] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:31:05.379 [2024-11-26 19:27:18.002282] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:31:05.948 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:05.948 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@868 -- # return 0 00:31:05.948 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:31:05.948 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:05.948 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 [2024-11-26 19:27:18.605111] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 Malloc0 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 Delay0 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 [2024-11-26 19:27:18.697024] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:06.209 19:27:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:31:06.209 [2024-11-26 19:27:18.818293] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:31:08.756 Initializing NVMe Controllers 00:31:08.756 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:31:08.756 controller IO queue size 128 less than required 00:31:08.756 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:31:08.756 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:31:08.756 Initialization complete. Launching workers. 00:31:08.756 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 28650 00:31:08.756 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 28707, failed to submit 66 00:31:08.756 success 28650, unsuccessful 57, failed 0 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@516 -- # nvmfcleanup 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:08.756 rmmod nvme_tcp 00:31:08.756 rmmod nvme_fabrics 00:31:08.756 rmmod nvme_keyring 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@517 -- # '[' -n 4072458 ']' 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@518 -- # killprocess 4072458 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@954 -- # '[' -z 4072458 ']' 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@958 -- # kill -0 4072458 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@959 -- # uname 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:08.756 19:27:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4072458 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4072458' 00:31:08.756 killing process with pid 4072458 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@973 -- # kill 4072458 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@978 -- # wait 4072458 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@791 -- # iptables-save 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@791 -- # iptables-restore 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:08.756 19:27:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:10.668 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:10.668 00:31:10.668 real 0m14.400s 00:31:10.668 user 0m11.045s 00:31:10.668 sys 0m7.696s 00:31:10.668 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:10.668 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:31:10.668 ************************************ 00:31:10.668 END TEST nvmf_abort 00:31:10.668 ************************************ 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:31:10.930 ************************************ 00:31:10.930 START TEST nvmf_ns_hotplug_stress 00:31:10.930 ************************************ 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:31:10.930 * Looking for test storage... 00:31:10.930 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1693 -- # lcov --version 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:31:10.930 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:10.930 --rc genhtml_branch_coverage=1 00:31:10.930 --rc genhtml_function_coverage=1 00:31:10.930 --rc genhtml_legend=1 00:31:10.930 --rc geninfo_all_blocks=1 00:31:10.930 --rc geninfo_unexecuted_blocks=1 00:31:10.930 00:31:10.930 ' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:31:10.930 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:10.930 --rc genhtml_branch_coverage=1 00:31:10.930 --rc genhtml_function_coverage=1 00:31:10.930 --rc genhtml_legend=1 00:31:10.930 --rc geninfo_all_blocks=1 00:31:10.930 --rc geninfo_unexecuted_blocks=1 00:31:10.930 00:31:10.930 ' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:31:10.930 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:10.930 --rc genhtml_branch_coverage=1 00:31:10.930 --rc genhtml_function_coverage=1 00:31:10.930 --rc genhtml_legend=1 00:31:10.930 --rc geninfo_all_blocks=1 00:31:10.930 --rc geninfo_unexecuted_blocks=1 00:31:10.930 00:31:10.930 ' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:31:10.930 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:10.930 --rc genhtml_branch_coverage=1 00:31:10.930 --rc genhtml_function_coverage=1 00:31:10.930 --rc genhtml_legend=1 00:31:10.930 --rc geninfo_all_blocks=1 00:31:10.930 --rc geninfo_unexecuted_blocks=1 00:31:10.930 00:31:10.930 ' 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:31:10.930 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # prepare_net_devs 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # local -g is_hw=no 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # remove_spdk_ns 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:31:10.931 19:27:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:31:19.075 Found 0000:31:00.0 (0x8086 - 0x159b) 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:19.075 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:31:19.076 Found 0000:31:00.1 (0x8086 - 0x159b) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:31:19.076 Found net devices under 0000:31:00.0: cvl_0_0 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:31:19.076 Found net devices under 0000:31:00.1: cvl_0_1 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # is_hw=yes 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:19.076 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:19.337 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:19.338 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:19.338 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.644 ms 00:31:19.338 00:31:19.338 --- 10.0.0.2 ping statistics --- 00:31:19.338 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:19.338 rtt min/avg/max/mdev = 0.644/0.644/0.644/0.000 ms 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:19.338 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:19.338 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.314 ms 00:31:19.338 00:31:19.338 --- 10.0.0.1 ping statistics --- 00:31:19.338 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:19.338 rtt min/avg/max/mdev = 0.314/0.314/0.314/0.000 ms 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@450 -- # return 0 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:31:19.338 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # nvmfpid=4077758 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # waitforlisten 4077758 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # '[' -z 4077758 ']' 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:19.600 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:19.600 19:27:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:31:19.600 [2024-11-26 19:27:32.038308] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:31:19.600 [2024-11-26 19:27:32.039875] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:31:19.600 [2024-11-26 19:27:32.039948] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:19.600 [2024-11-26 19:27:32.148441] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:31:19.600 [2024-11-26 19:27:32.199279] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:19.600 [2024-11-26 19:27:32.199336] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:19.600 [2024-11-26 19:27:32.199345] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:19.600 [2024-11-26 19:27:32.199353] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:19.600 [2024-11-26 19:27:32.199360] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:19.600 [2024-11-26 19:27:32.201220] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:31:19.600 [2024-11-26 19:27:32.201385] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:19.600 [2024-11-26 19:27:32.201385] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:31:19.861 [2024-11-26 19:27:32.278260] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:31:19.861 [2024-11-26 19:27:32.278307] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:31:19.861 [2024-11-26 19:27:32.278887] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:31:19.861 [2024-11-26 19:27:32.279188] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@868 -- # return 0 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:31:20.432 19:27:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:31:20.432 [2024-11-26 19:27:33.026299] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:20.432 19:27:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:31:20.692 19:27:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:31:20.953 [2024-11-26 19:27:33.406997] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:20.953 19:27:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:31:21.214 19:27:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:31:21.214 Malloc0 00:31:21.214 19:27:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:31:21.477 Delay0 00:31:21.477 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:21.737 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:31:21.737 NULL1 00:31:21.998 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:31:21.998 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=4078189 00:31:21.998 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:21.998 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:31:21.998 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:22.258 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:22.519 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:31:22.519 19:27:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:31:22.519 true 00:31:22.519 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:22.519 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:22.779 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:23.040 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:31:23.040 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:31:23.040 true 00:31:23.040 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:23.040 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:23.300 19:27:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:23.559 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:31:23.559 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:31:23.559 true 00:31:23.819 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:23.819 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:23.819 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:24.079 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:31:24.079 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:31:24.339 true 00:31:24.339 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:24.339 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:24.601 19:27:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:24.601 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:31:24.601 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:31:24.863 true 00:31:24.863 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:24.863 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:25.123 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:25.123 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:31:25.123 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:31:25.383 true 00:31:25.383 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:25.383 19:27:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:25.643 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:25.903 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:31:25.904 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:31:25.904 true 00:31:25.904 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:25.904 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:26.165 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:26.425 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:31:26.425 19:27:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:31:26.425 true 00:31:26.687 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:26.687 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:26.687 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:26.947 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:31:26.947 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:31:27.207 true 00:31:27.207 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:27.207 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:27.207 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:27.469 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:31:27.469 19:27:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:31:27.729 true 00:31:27.729 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:27.729 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:27.991 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:27.991 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:31:27.991 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:31:28.251 true 00:31:28.252 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:28.252 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:28.512 19:27:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:28.512 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:31:28.512 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:31:28.773 true 00:31:28.773 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:28.773 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:29.034 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:29.034 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:31:29.034 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:31:29.294 true 00:31:29.294 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:29.294 19:27:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:29.554 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:29.816 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:31:29.816 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:31:29.816 true 00:31:29.816 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:29.816 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:30.075 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:30.335 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:31:30.335 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:31:30.336 true 00:31:30.336 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:30.336 19:27:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:30.596 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:30.856 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:31:30.856 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:31:30.856 true 00:31:31.117 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:31.117 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:31.117 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:31.377 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:31:31.377 19:27:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:31:31.637 true 00:31:31.637 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:31.637 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:31.637 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:31.897 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:31:31.897 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:31:32.158 true 00:31:32.158 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:32.158 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:32.417 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:32.417 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:31:32.417 19:27:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:31:32.677 true 00:31:32.677 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:32.677 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:32.936 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:32.936 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:31:32.936 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:31:33.196 true 00:31:33.196 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:33.196 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:33.457 19:27:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:33.718 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:31:33.718 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:31:33.718 true 00:31:33.718 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:33.718 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:33.979 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:34.239 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:31:34.239 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:31:34.239 true 00:31:34.239 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:34.239 19:27:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:34.499 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:34.760 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:31:34.760 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:31:34.760 true 00:31:35.022 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:35.022 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:35.022 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:35.282 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:31:35.282 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:31:35.543 true 00:31:35.543 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:35.543 19:27:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:35.543 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:35.804 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:31:35.804 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:31:36.064 true 00:31:36.064 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:36.064 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:36.324 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:36.324 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:31:36.324 19:27:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:31:36.585 true 00:31:36.585 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:36.585 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:36.845 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:36.845 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:31:36.845 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:31:37.106 true 00:31:37.106 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:37.106 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:37.366 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:37.366 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:31:37.367 19:27:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:31:37.627 true 00:31:37.627 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:37.627 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:37.888 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:38.149 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:31:38.149 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:31:38.149 true 00:31:38.149 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:38.149 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:38.409 19:27:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:38.672 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1030 00:31:38.672 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1030 00:31:38.672 true 00:31:38.672 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:38.672 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:38.934 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:39.196 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1031 00:31:39.196 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1031 00:31:39.196 true 00:31:39.457 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:39.457 19:27:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:39.457 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:39.718 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1032 00:31:39.718 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1032 00:31:39.980 true 00:31:39.980 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:39.980 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:39.980 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:40.259 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1033 00:31:40.259 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1033 00:31:40.519 true 00:31:40.519 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:40.519 19:27:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:40.780 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:40.780 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1034 00:31:40.780 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1034 00:31:41.040 true 00:31:41.040 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:41.040 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:41.302 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:41.302 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1035 00:31:41.302 19:27:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1035 00:31:41.562 true 00:31:41.562 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:41.562 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:41.822 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:42.082 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1036 00:31:42.082 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1036 00:31:42.082 true 00:31:42.082 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:42.082 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:42.343 19:27:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:42.602 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1037 00:31:42.602 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1037 00:31:42.602 true 00:31:42.602 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:42.602 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:42.861 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:43.122 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1038 00:31:43.122 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1038 00:31:43.122 true 00:31:43.381 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:43.381 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:43.381 19:27:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:43.641 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1039 00:31:43.641 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1039 00:31:43.901 true 00:31:43.901 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:43.901 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:43.901 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:44.162 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1040 00:31:44.162 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1040 00:31:44.422 true 00:31:44.422 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:44.422 19:27:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:44.682 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:44.683 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1041 00:31:44.683 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1041 00:31:44.943 true 00:31:44.943 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:44.943 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:45.203 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:45.203 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1042 00:31:45.203 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1042 00:31:45.463 true 00:31:45.463 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:45.463 19:27:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:45.723 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:45.984 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1043 00:31:45.984 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1043 00:31:45.984 true 00:31:45.984 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:45.984 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:46.244 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:46.506 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1044 00:31:46.506 19:27:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1044 00:31:46.506 true 00:31:46.506 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:46.506 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:46.777 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:47.040 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1045 00:31:47.040 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1045 00:31:47.040 true 00:31:47.040 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:47.040 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:47.300 19:27:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:47.561 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1046 00:31:47.561 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1046 00:31:47.561 true 00:31:47.822 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:47.822 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:47.822 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:48.083 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1047 00:31:48.083 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1047 00:31:48.343 true 00:31:48.343 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:48.343 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:48.343 19:28:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:48.604 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1048 00:31:48.604 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1048 00:31:48.865 true 00:31:48.865 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:48.865 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:49.126 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:49.126 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1049 00:31:49.127 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1049 00:31:49.388 true 00:31:49.388 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:49.388 19:28:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:49.648 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:49.648 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1050 00:31:49.648 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1050 00:31:49.908 true 00:31:49.908 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:49.908 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:50.169 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:50.429 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1051 00:31:50.429 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1051 00:31:50.429 true 00:31:50.429 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:50.429 19:28:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:50.690 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:50.951 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1052 00:31:50.951 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1052 00:31:50.951 true 00:31:50.951 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:50.951 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:51.213 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:51.474 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1053 00:31:51.474 19:28:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1053 00:31:51.474 true 00:31:51.734 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:51.734 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:51.734 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:51.995 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1054 00:31:51.995 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1054 00:31:52.256 true 00:31:52.256 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:52.256 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:52.256 19:28:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:31:52.517 Initializing NVMe Controllers 00:31:52.517 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:31:52.517 Controller IO queue size 128, less than required. 00:31:52.517 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:31:52.517 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:31:52.517 Initialization complete. Launching workers. 00:31:52.517 ======================================================== 00:31:52.517 Latency(us) 00:31:52.517 Device Information : IOPS MiB/s Average min max 00:31:52.517 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 29272.16 14.29 4372.71 1477.68 11309.19 00:31:52.517 ======================================================== 00:31:52.517 Total : 29272.16 14.29 4372.71 1477.68 11309.19 00:31:52.517 00:31:52.517 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1055 00:31:52.517 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1055 00:31:52.778 true 00:31:52.778 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4078189 00:31:52.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (4078189) - No such process 00:31:52.778 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 4078189 00:31:52.778 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:52.778 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:53.038 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:31:53.038 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:31:53.038 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:31:53.038 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:53.038 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:31:53.299 null0 00:31:53.299 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:53.299 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:53.299 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:31:53.299 null1 00:31:53.299 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:53.299 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:53.299 19:28:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:31:53.561 null2 00:31:53.561 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:53.561 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:53.561 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:31:53.832 null3 00:31:53.832 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:53.832 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:53.833 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:31:53.833 null4 00:31:53.833 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:53.833 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:53.833 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:31:54.094 null5 00:31:54.094 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:54.094 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:54.094 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:31:54.356 null6 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:31:54.356 null7 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:31:54.356 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 4084382 4084384 4084386 4084388 4084390 4084392 4084394 4084396 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.357 19:28:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:54.618 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.879 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:54.880 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:54.880 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:54.880 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:54.880 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:54.880 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.141 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:55.402 19:28:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:55.402 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:55.402 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.663 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:55.664 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:55.664 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.664 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.664 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:55.664 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:55.925 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:56.186 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:56.187 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.449 19:28:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:56.449 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:56.449 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:56.449 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.449 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.449 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:56.449 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.711 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:56.973 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:57.233 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:57.493 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:57.493 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.493 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.493 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:57.494 19:28:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:31:57.494 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:57.755 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.016 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.276 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # nvmfcleanup 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:58.537 rmmod nvme_tcp 00:31:58.537 rmmod nvme_fabrics 00:31:58.537 rmmod nvme_keyring 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@517 -- # '[' -n 4077758 ']' 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # killprocess 4077758 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # '[' -z 4077758 ']' 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@958 -- # kill -0 4077758 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # uname 00:31:58.537 19:28:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:58.537 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4077758 00:31:58.537 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:31:58.537 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:31:58.537 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4077758' 00:31:58.537 killing process with pid 4077758 00:31:58.537 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@973 -- # kill 4077758 00:31:58.537 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@978 -- # wait 4077758 00:31:58.798 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-save 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-restore 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:58.799 19:28:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:00.713 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:00.713 00:32:00.713 real 0m49.956s 00:32:00.713 user 3m4.947s 00:32:00.713 sys 0m22.613s 00:32:00.713 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:00.713 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:32:00.713 ************************************ 00:32:00.713 END TEST nvmf_ns_hotplug_stress 00:32:00.713 ************************************ 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:32:00.974 ************************************ 00:32:00.974 START TEST nvmf_delete_subsystem 00:32:00.974 ************************************ 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:32:00.974 * Looking for test storage... 00:32:00.974 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1693 -- # lcov --version 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:00.974 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:32:00.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:00.975 --rc genhtml_branch_coverage=1 00:32:00.975 --rc genhtml_function_coverage=1 00:32:00.975 --rc genhtml_legend=1 00:32:00.975 --rc geninfo_all_blocks=1 00:32:00.975 --rc geninfo_unexecuted_blocks=1 00:32:00.975 00:32:00.975 ' 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:32:00.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:00.975 --rc genhtml_branch_coverage=1 00:32:00.975 --rc genhtml_function_coverage=1 00:32:00.975 --rc genhtml_legend=1 00:32:00.975 --rc geninfo_all_blocks=1 00:32:00.975 --rc geninfo_unexecuted_blocks=1 00:32:00.975 00:32:00.975 ' 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:32:00.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:00.975 --rc genhtml_branch_coverage=1 00:32:00.975 --rc genhtml_function_coverage=1 00:32:00.975 --rc genhtml_legend=1 00:32:00.975 --rc geninfo_all_blocks=1 00:32:00.975 --rc geninfo_unexecuted_blocks=1 00:32:00.975 00:32:00.975 ' 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:32:00.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:00.975 --rc genhtml_branch_coverage=1 00:32:00.975 --rc genhtml_function_coverage=1 00:32:00.975 --rc genhtml_legend=1 00:32:00.975 --rc geninfo_all_blocks=1 00:32:00.975 --rc geninfo_unexecuted_blocks=1 00:32:00.975 00:32:00.975 ' 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:00.975 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:32:01.237 19:28:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:09.381 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:32:09.382 Found 0000:31:00.0 (0x8086 - 0x159b) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:32:09.382 Found 0000:31:00.1 (0x8086 - 0x159b) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:32:09.382 Found net devices under 0000:31:00.0: cvl_0_0 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:32:09.382 Found net devices under 0000:31:00.1: cvl_0_1 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # is_hw=yes 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:09.382 19:28:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:09.643 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:09.643 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:09.643 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:09.643 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:09.643 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:09.643 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.568 ms 00:32:09.643 00:32:09.643 --- 10.0.0.2 ping statistics --- 00:32:09.643 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:09.643 rtt min/avg/max/mdev = 0.568/0.568/0.568/0.000 ms 00:32:09.643 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:09.643 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:09.643 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.274 ms 00:32:09.643 00:32:09.643 --- 10.0.0.1 ping statistics --- 00:32:09.643 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:09.643 rtt min/avg/max/mdev = 0.274/0.274/0.274/0.000 ms 00:32:09.643 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:09.643 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@450 -- # return 0 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # nvmfpid=4090204 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # waitforlisten 4090204 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # '[' -z 4090204 ']' 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:09.644 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:09.644 19:28:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:09.644 [2024-11-26 19:28:22.250726] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:32:09.644 [2024-11-26 19:28:22.252090] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:32:09.644 [2024-11-26 19:28:22.252156] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:09.904 [2024-11-26 19:28:22.343834] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:32:09.904 [2024-11-26 19:28:22.384136] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:09.904 [2024-11-26 19:28:22.384168] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:09.904 [2024-11-26 19:28:22.384176] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:09.904 [2024-11-26 19:28:22.384183] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:09.904 [2024-11-26 19:28:22.384190] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:09.904 [2024-11-26 19:28:22.385419] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:09.904 [2024-11-26 19:28:22.385422] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:09.904 [2024-11-26 19:28:22.442071] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:32:09.904 [2024-11-26 19:28:22.442588] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:32:09.904 [2024-11-26 19:28:22.442951] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:32:10.529 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:10.529 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@868 -- # return 0 00:32:10.529 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:10.529 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:10.529 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:10.530 [2024-11-26 19:28:23.090336] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:10.530 [2024-11-26 19:28:23.118691] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:10.530 NULL1 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:10.530 Delay0 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:10.530 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:10.842 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:10.842 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=4090244 00:32:10.842 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:32:10.842 19:28:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:32:10.842 [2024-11-26 19:28:23.215811] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:32:12.850 19:28:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:32:12.850 19:28:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:12.850 19:28:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 starting I/O failed: -6 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 starting I/O failed: -6 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 starting I/O failed: -6 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 starting I/O failed: -6 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 starting I/O failed: -6 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 starting I/O failed: -6 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 starting I/O failed: -6 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Write completed with error (sct=0, sc=8) 00:32:12.850 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 [2024-11-26 19:28:25.387527] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7a2f00 is same with the state(6) to be set 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 starting I/O failed: -6 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 [2024-11-26 19:28:25.390602] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ff27000d4b0 is same with the state(6) to be set 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Write completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.851 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Write completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Write completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Write completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Write completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Write completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:12.852 Read completed with error (sct=0, sc=8) 00:32:13.794 [2024-11-26 19:28:26.353887] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7a45f0 is same with the state(6) to be set 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 [2024-11-26 19:28:26.390829] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7a30e0 is same with the state(6) to be set 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 [2024-11-26 19:28:26.391208] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7a34a0 is same with the state(6) to be set 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 [2024-11-26 19:28:26.393325] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ff27000d7e0 is same with the state(6) to be set 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Write completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 Read completed with error (sct=0, sc=8) 00:32:13.794 [2024-11-26 19:28:26.393418] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ff27000d020 is same with the state(6) to be set 00:32:13.794 Initializing NVMe Controllers 00:32:13.794 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:13.794 Controller IO queue size 128, less than required. 00:32:13.794 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:13.794 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:32:13.794 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:32:13.794 Initialization complete. Launching workers. 00:32:13.794 ======================================================== 00:32:13.794 Latency(us) 00:32:13.794 Device Information : IOPS MiB/s Average min max 00:32:13.794 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 168.79 0.08 897138.93 249.02 1007005.32 00:32:13.794 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 162.81 0.08 955388.69 278.82 2002145.27 00:32:13.794 ======================================================== 00:32:13.794 Total : 331.60 0.16 925739.04 249.02 2002145.27 00:32:13.794 00:32:13.794 [2024-11-26 19:28:26.394115] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x7a45f0 (9): Bad file descriptor 00:32:13.794 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:32:13.794 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:13.794 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:32:13.794 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 4090244 00:32:13.794 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:32:14.366 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 4090244 00:32:14.367 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (4090244) - No such process 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 4090244 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # local es=0 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@654 -- # valid_exec_arg wait 4090244 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@640 -- # local arg=wait 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # type -t wait 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # wait 4090244 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # es=1 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:14.367 [2024-11-26 19:28:26.926404] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=4090988 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:14.367 19:28:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:32:14.627 [2024-11-26 19:28:26.997227] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:32:14.888 19:28:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:32:14.888 19:28:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:14.888 19:28:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:32:15.458 19:28:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:32:15.458 19:28:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:15.458 19:28:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:32:16.028 19:28:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:32:16.028 19:28:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:16.028 19:28:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:32:16.598 19:28:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:32:16.598 19:28:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:16.598 19:28:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:32:16.859 19:28:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:32:16.859 19:28:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:16.859 19:28:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:32:17.430 19:28:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:32:17.430 19:28:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:17.430 19:28:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:32:17.430 Initializing NVMe Controllers 00:32:17.430 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:17.430 Controller IO queue size 128, less than required. 00:32:17.430 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:17.430 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:32:17.430 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:32:17.430 Initialization complete. Launching workers. 00:32:17.430 ======================================================== 00:32:17.430 Latency(us) 00:32:17.430 Device Information : IOPS MiB/s Average min max 00:32:17.430 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1002297.47 1000227.61 1005639.33 00:32:17.430 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1004116.93 1000376.90 1010690.82 00:32:17.430 ======================================================== 00:32:17.430 Total : 256.00 0.12 1003207.20 1000227.61 1010690.82 00:32:17.430 00:32:17.430 [2024-11-26 19:28:30.046939] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13b1a10 is same with the state(6) to be set 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4090988 00:32:18.002 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (4090988) - No such process 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 4090988 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:18.002 rmmod nvme_tcp 00:32:18.002 rmmod nvme_fabrics 00:32:18.002 rmmod nvme_keyring 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@517 -- # '[' -n 4090204 ']' 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # killprocess 4090204 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # '[' -z 4090204 ']' 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@958 -- # kill -0 4090204 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # uname 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4090204 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4090204' 00:32:18.002 killing process with pid 4090204 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@973 -- # kill 4090204 00:32:18.002 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@978 -- # wait 4090204 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-save 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-restore 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:18.263 19:28:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:20.810 00:32:20.810 real 0m19.441s 00:32:20.810 user 0m26.723s 00:32:20.810 sys 0m8.251s 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:32:20.810 ************************************ 00:32:20.810 END TEST nvmf_delete_subsystem 00:32:20.810 ************************************ 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:32:20.810 ************************************ 00:32:20.810 START TEST nvmf_host_management 00:32:20.810 ************************************ 00:32:20.810 19:28:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:32:20.810 * Looking for test storage... 00:32:20.810 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1693 -- # lcov --version 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:32:20.810 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:32:20.811 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:20.811 --rc genhtml_branch_coverage=1 00:32:20.811 --rc genhtml_function_coverage=1 00:32:20.811 --rc genhtml_legend=1 00:32:20.811 --rc geninfo_all_blocks=1 00:32:20.811 --rc geninfo_unexecuted_blocks=1 00:32:20.811 00:32:20.811 ' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:32:20.811 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:20.811 --rc genhtml_branch_coverage=1 00:32:20.811 --rc genhtml_function_coverage=1 00:32:20.811 --rc genhtml_legend=1 00:32:20.811 --rc geninfo_all_blocks=1 00:32:20.811 --rc geninfo_unexecuted_blocks=1 00:32:20.811 00:32:20.811 ' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:32:20.811 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:20.811 --rc genhtml_branch_coverage=1 00:32:20.811 --rc genhtml_function_coverage=1 00:32:20.811 --rc genhtml_legend=1 00:32:20.811 --rc geninfo_all_blocks=1 00:32:20.811 --rc geninfo_unexecuted_blocks=1 00:32:20.811 00:32:20.811 ' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:32:20.811 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:20.811 --rc genhtml_branch_coverage=1 00:32:20.811 --rc genhtml_function_coverage=1 00:32:20.811 --rc genhtml_legend=1 00:32:20.811 --rc geninfo_all_blocks=1 00:32:20.811 --rc geninfo_unexecuted_blocks=1 00:32:20.811 00:32:20.811 ' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:20.811 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:20.812 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:20.812 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:20.812 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:32:20.812 19:28:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:28.959 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:32:28.960 Found 0000:31:00.0 (0x8086 - 0x159b) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:32:28.960 Found 0000:31:00.1 (0x8086 - 0x159b) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:32:28.960 Found net devices under 0000:31:00.0: cvl_0_0 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:32:28.960 Found net devices under 0000:31:00.1: cvl_0_1 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # is_hw=yes 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:28.960 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:28.960 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.478 ms 00:32:28.960 00:32:28.960 --- 10.0.0.2 ping statistics --- 00:32:28.960 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:28.960 rtt min/avg/max/mdev = 0.478/0.478/0.478/0.000 ms 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:28.960 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:28.960 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.324 ms 00:32:28.960 00:32:28.960 --- 10.0.0.1 ping statistics --- 00:32:28.960 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:28.960 rtt min/avg/max/mdev = 0.324/0.324/0.324/0.000 ms 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@450 -- # return 0 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:28.960 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@509 -- # nvmfpid=4096422 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@510 -- # waitforlisten 4096422 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1E 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 4096422 ']' 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:29.223 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:29.223 19:28:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:29.223 [2024-11-26 19:28:41.683253] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:32:29.223 [2024-11-26 19:28:41.684409] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:32:29.223 [2024-11-26 19:28:41.684463] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:29.223 [2024-11-26 19:28:41.793154] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:32:29.484 [2024-11-26 19:28:41.846501] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:29.484 [2024-11-26 19:28:41.846554] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:29.484 [2024-11-26 19:28:41.846562] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:29.484 [2024-11-26 19:28:41.846569] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:29.484 [2024-11-26 19:28:41.846576] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:29.484 [2024-11-26 19:28:41.848565] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:29.484 [2024-11-26 19:28:41.848611] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:29.484 [2024-11-26 19:28:41.848777] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:29.484 [2024-11-26 19:28:41.848778] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:32:29.484 [2024-11-26 19:28:41.923946] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:32:29.484 [2024-11-26 19:28:41.924634] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:32:29.484 [2024-11-26 19:28:41.925522] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:32:29.484 [2024-11-26 19:28:41.925578] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:32:29.484 [2024-11-26 19:28:41.925908] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:30.056 [2024-11-26 19:28:42.541662] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:30.056 Malloc0 00:32:30.056 [2024-11-26 19:28:42.629776] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:30.056 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=4096649 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 4096649 /var/tmp/bdevperf.sock 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 4096649 ']' 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:32:30.316 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:32:30.316 { 00:32:30.316 "params": { 00:32:30.316 "name": "Nvme$subsystem", 00:32:30.316 "trtype": "$TEST_TRANSPORT", 00:32:30.316 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:30.316 "adrfam": "ipv4", 00:32:30.316 "trsvcid": "$NVMF_PORT", 00:32:30.316 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:30.316 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:30.316 "hdgst": ${hdgst:-false}, 00:32:30.316 "ddgst": ${ddgst:-false} 00:32:30.316 }, 00:32:30.316 "method": "bdev_nvme_attach_controller" 00:32:30.316 } 00:32:30.316 EOF 00:32:30.316 )") 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:32:30.316 19:28:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:32:30.316 "params": { 00:32:30.316 "name": "Nvme0", 00:32:30.316 "trtype": "tcp", 00:32:30.316 "traddr": "10.0.0.2", 00:32:30.316 "adrfam": "ipv4", 00:32:30.316 "trsvcid": "4420", 00:32:30.316 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:32:30.316 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:32:30.316 "hdgst": false, 00:32:30.316 "ddgst": false 00:32:30.317 }, 00:32:30.317 "method": "bdev_nvme_attach_controller" 00:32:30.317 }' 00:32:30.317 [2024-11-26 19:28:42.733256] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:32:30.317 [2024-11-26 19:28:42.733309] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4096649 ] 00:32:30.317 [2024-11-26 19:28:42.812714] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:30.317 [2024-11-26 19:28:42.849078] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:30.578 Running I/O for 10 seconds... 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=963 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@58 -- # '[' 963 -ge 100 ']' 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@60 -- # break 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:31.150 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:31.150 [2024-11-26 19:28:43.593564] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:32:31.150 [2024-11-26 19:28:43.593603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.593614] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:32:31.150 [2024-11-26 19:28:43.593621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.593630] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:32:31.150 [2024-11-26 19:28:43.593638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.593646] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:32:31.150 [2024-11-26 19:28:43.593653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.593660] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1e1ab10 is same with the state(6) to be set 00:32:31.150 [2024-11-26 19:28:43.596350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:127616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:127744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:127872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:128000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:128128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:128256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:128384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:128512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:128640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:128768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:128896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:129024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:129152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:129280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:129408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:129536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:129664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:129792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:129920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:130048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:130176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.150 [2024-11-26 19:28:43.596719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.150 [2024-11-26 19:28:43.596728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:130304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:130432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:130560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:130688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:130816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:130944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:1024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.596987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:1152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.596995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:1280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:1408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:1536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:1664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:1792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:2048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:2176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:2304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:2432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:2560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:2688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:2816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:2944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:3072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:3200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:3328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:3456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:3584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:3712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:3840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:3968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:4096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.151 [2024-11-26 19:28:43.597380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.151 [2024-11-26 19:28:43.597389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:4224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.152 [2024-11-26 19:28:43.597397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.152 [2024-11-26 19:28:43.597406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:4352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.152 [2024-11-26 19:28:43.597414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.152 [2024-11-26 19:28:43.597423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:4480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.152 [2024-11-26 19:28:43.597430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.152 [2024-11-26 19:28:43.597439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:4608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:31.152 [2024-11-26 19:28:43.597446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.152 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:31.152 [2024-11-26 19:28:43.598693] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:32:31.152 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:32:31.152 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:31.152 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:31.152 task offset: 127616 on job bdev=Nvme0n1 fails 00:32:31.152 00:32:31.152 Latency(us) 00:32:31.152 [2024-11-26T18:28:43.777Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:31.152 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:32:31.152 Job: Nvme0n1 ended in about 0.61 seconds with error 00:32:31.152 Verification LBA range: start 0x0 length 0x400 00:32:31.152 Nvme0n1 : 0.61 1644.04 102.75 105.53 0.00 35713.81 1508.69 31457.28 00:32:31.152 [2024-11-26T18:28:43.777Z] =================================================================================================================== 00:32:31.152 [2024-11-26T18:28:43.777Z] Total : 1644.04 102.75 105.53 0.00 35713.81 1508.69 31457.28 00:32:31.152 [2024-11-26 19:28:43.600689] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:32:31.152 [2024-11-26 19:28:43.600712] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1e1ab10 (9): Bad file descriptor 00:32:31.152 [2024-11-26 19:28:43.601992] ctrlr.c: 825:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode0' does not allow host 'nqn.2016-06.io.spdk:host0' 00:32:31.152 [2024-11-26 19:28:43.602066] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:3 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:32:31.152 [2024-11-26 19:28:43.602086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND SPECIFIC (01/84) qid:0 cid:3 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:31.152 [2024-11-26 19:28:43.602098] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode0 00:32:31.152 [2024-11-26 19:28:43.602106] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 132 00:32:31.152 [2024-11-26 19:28:43.602113] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:32:31.152 [2024-11-26 19:28:43.602120] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1e1ab10 00:32:31.152 [2024-11-26 19:28:43.602139] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1e1ab10 (9): Bad file descriptor 00:32:31.152 [2024-11-26 19:28:43.602151] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:32:31.152 [2024-11-26 19:28:43.602159] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:32:31.152 [2024-11-26 19:28:43.602167] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:32:31.152 [2024-11-26 19:28:43.602176] bdev_nvme.c:2280:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:32:31.152 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:31.152 19:28:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:32:32.092 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 4096649 00:32:32.092 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (4096649) - No such process 00:32:32.092 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # true 00:32:32.092 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:32:32.092 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:32:32.092 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:32:32.093 { 00:32:32.093 "params": { 00:32:32.093 "name": "Nvme$subsystem", 00:32:32.093 "trtype": "$TEST_TRANSPORT", 00:32:32.093 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:32.093 "adrfam": "ipv4", 00:32:32.093 "trsvcid": "$NVMF_PORT", 00:32:32.093 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:32.093 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:32.093 "hdgst": ${hdgst:-false}, 00:32:32.093 "ddgst": ${ddgst:-false} 00:32:32.093 }, 00:32:32.093 "method": "bdev_nvme_attach_controller" 00:32:32.093 } 00:32:32.093 EOF 00:32:32.093 )") 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:32:32.093 19:28:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:32:32.093 "params": { 00:32:32.093 "name": "Nvme0", 00:32:32.093 "trtype": "tcp", 00:32:32.093 "traddr": "10.0.0.2", 00:32:32.093 "adrfam": "ipv4", 00:32:32.093 "trsvcid": "4420", 00:32:32.093 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:32:32.093 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:32:32.093 "hdgst": false, 00:32:32.093 "ddgst": false 00:32:32.093 }, 00:32:32.093 "method": "bdev_nvme_attach_controller" 00:32:32.093 }' 00:32:32.093 [2024-11-26 19:28:44.669121] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:32:32.093 [2024-11-26 19:28:44.669176] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4097004 ] 00:32:32.355 [2024-11-26 19:28:44.746947] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:32.355 [2024-11-26 19:28:44.783605] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:32.355 Running I/O for 1 seconds... 00:32:33.740 1679.00 IOPS, 104.94 MiB/s 00:32:33.740 Latency(us) 00:32:33.740 [2024-11-26T18:28:46.366Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:33.741 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:32:33.741 Verification LBA range: start 0x0 length 0x400 00:32:33.741 Nvme0n1 : 1.01 1724.11 107.76 0.00 0.00 36400.07 2498.56 38229.33 00:32:33.741 [2024-11-26T18:28:46.366Z] =================================================================================================================== 00:32:33.741 [2024-11-26T18:28:46.366Z] Total : 1724.11 107.76 0.00 0.00 36400.07 2498.56 38229.33 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:33.741 rmmod nvme_tcp 00:32:33.741 rmmod nvme_fabrics 00:32:33.741 rmmod nvme_keyring 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@517 -- # '[' -n 4096422 ']' 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@518 -- # killprocess 4096422 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@954 -- # '[' -z 4096422 ']' 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@958 -- # kill -0 4096422 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@959 -- # uname 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4096422 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4096422' 00:32:33.741 killing process with pid 4096422 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@973 -- # kill 4096422 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@978 -- # wait 4096422 00:32:33.741 [2024-11-26 19:28:46.291474] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-save 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-restore 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:33.741 19:28:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:32:36.289 00:32:36.289 real 0m15.485s 00:32:36.289 user 0m18.808s 00:32:36.289 sys 0m8.174s 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:32:36.289 ************************************ 00:32:36.289 END TEST nvmf_host_management 00:32:36.289 ************************************ 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:32:36.289 ************************************ 00:32:36.289 START TEST nvmf_lvol 00:32:36.289 ************************************ 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:32:36.289 * Looking for test storage... 00:32:36.289 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1693 -- # lcov --version 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:36.289 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:32:36.290 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:36.290 --rc genhtml_branch_coverage=1 00:32:36.290 --rc genhtml_function_coverage=1 00:32:36.290 --rc genhtml_legend=1 00:32:36.290 --rc geninfo_all_blocks=1 00:32:36.290 --rc geninfo_unexecuted_blocks=1 00:32:36.290 00:32:36.290 ' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:32:36.290 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:36.290 --rc genhtml_branch_coverage=1 00:32:36.290 --rc genhtml_function_coverage=1 00:32:36.290 --rc genhtml_legend=1 00:32:36.290 --rc geninfo_all_blocks=1 00:32:36.290 --rc geninfo_unexecuted_blocks=1 00:32:36.290 00:32:36.290 ' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:32:36.290 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:36.290 --rc genhtml_branch_coverage=1 00:32:36.290 --rc genhtml_function_coverage=1 00:32:36.290 --rc genhtml_legend=1 00:32:36.290 --rc geninfo_all_blocks=1 00:32:36.290 --rc geninfo_unexecuted_blocks=1 00:32:36.290 00:32:36.290 ' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:32:36.290 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:36.290 --rc genhtml_branch_coverage=1 00:32:36.290 --rc genhtml_function_coverage=1 00:32:36.290 --rc genhtml_legend=1 00:32:36.290 --rc geninfo_all_blocks=1 00:32:36.290 --rc geninfo_unexecuted_blocks=1 00:32:36.290 00:32:36.290 ' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:32:36.290 19:28:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:32:44.428 Found 0000:31:00.0 (0x8086 - 0x159b) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:32:44.428 Found 0000:31:00.1 (0x8086 - 0x159b) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:44.428 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:32:44.429 Found net devices under 0000:31:00.0: cvl_0_0 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:32:44.429 Found net devices under 0000:31:00.1: cvl_0_1 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # is_hw=yes 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:44.429 19:28:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:44.429 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:44.429 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:44.429 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:44.429 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:44.690 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:44.690 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.657 ms 00:32:44.690 00:32:44.690 --- 10.0.0.2 ping statistics --- 00:32:44.690 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:44.690 rtt min/avg/max/mdev = 0.657/0.657/0.657/0.000 ms 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:44.690 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:44.690 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.309 ms 00:32:44.690 00:32:44.690 --- 10.0.0.1 ping statistics --- 00:32:44.690 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:44.690 rtt min/avg/max/mdev = 0.309/0.309/0.309/0.000 ms 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@450 -- # return 0 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@509 -- # nvmfpid=4102015 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@510 -- # waitforlisten 4102015 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@835 -- # '[' -z 4102015 ']' 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:44.690 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:32:44.690 19:28:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x7 00:32:44.690 [2024-11-26 19:28:57.263672] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:32:44.690 [2024-11-26 19:28:57.264860] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:32:44.690 [2024-11-26 19:28:57.264922] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:44.950 [2024-11-26 19:28:57.355649] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:32:44.950 [2024-11-26 19:28:57.397414] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:44.950 [2024-11-26 19:28:57.397452] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:44.950 [2024-11-26 19:28:57.397460] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:44.950 [2024-11-26 19:28:57.397467] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:44.950 [2024-11-26 19:28:57.397473] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:44.951 [2024-11-26 19:28:57.399123] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:44.951 [2024-11-26 19:28:57.399240] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:44.951 [2024-11-26 19:28:57.399242] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:44.951 [2024-11-26 19:28:57.456406] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:32:44.951 [2024-11-26 19:28:57.456927] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:32:44.951 [2024-11-26 19:28:57.457238] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:32:44.951 [2024-11-26 19:28:57.457409] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:32:45.522 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:45.522 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@868 -- # return 0 00:32:45.522 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:45.522 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:45.522 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:32:45.522 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:45.522 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:32:45.782 [2024-11-26 19:28:58.272157] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:45.782 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:32:46.043 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:32:46.043 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:32:46.303 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:32:46.303 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:32:46.303 19:28:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:32:46.563 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=5cac5208-0766-4222-a498-f981320ce0d9 00:32:46.563 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 5cac5208-0766-4222-a498-f981320ce0d9 lvol 20 00:32:46.563 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=a7c97f95-aa09-4a0d-8f41-6adf4a35eb7c 00:32:46.563 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:32:46.823 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 a7c97f95-aa09-4a0d-8f41-6adf4a35eb7c 00:32:47.083 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:32:47.083 [2024-11-26 19:28:59.647929] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:47.083 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:32:47.343 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=4102646 00:32:47.343 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:32:47.343 19:28:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:32:48.284 19:29:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot a7c97f95-aa09-4a0d-8f41-6adf4a35eb7c MY_SNAPSHOT 00:32:48.545 19:29:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=2d96c8db-87e7-4c72-a885-3264154282f1 00:32:48.545 19:29:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize a7c97f95-aa09-4a0d-8f41-6adf4a35eb7c 30 00:32:48.805 19:29:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone 2d96c8db-87e7-4c72-a885-3264154282f1 MY_CLONE 00:32:49.066 19:29:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=7aa46445-d53d-40b4-b938-e242c50fe6a9 00:32:49.066 19:29:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate 7aa46445-d53d-40b4-b938-e242c50fe6a9 00:32:49.327 19:29:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 4102646 00:32:57.543 Initializing NVMe Controllers 00:32:57.543 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:32:57.543 Controller IO queue size 128, less than required. 00:32:57.543 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:57.543 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:32:57.543 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:32:57.543 Initialization complete. Launching workers. 00:32:57.543 ======================================================== 00:32:57.543 Latency(us) 00:32:57.543 Device Information : IOPS MiB/s Average min max 00:32:57.543 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 15597.30 60.93 8207.26 1022.84 70678.90 00:32:57.543 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 12415.10 48.50 10312.98 2520.13 58599.65 00:32:57.543 ======================================================== 00:32:57.543 Total : 28012.40 109.42 9140.52 1022.84 70678.90 00:32:57.543 00:32:57.543 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:32:57.804 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete a7c97f95-aa09-4a0d-8f41-6adf4a35eb7c 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 5cac5208-0766-4222-a498-f981320ce0d9 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:58.066 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:58.066 rmmod nvme_tcp 00:32:58.328 rmmod nvme_fabrics 00:32:58.328 rmmod nvme_keyring 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@517 -- # '[' -n 4102015 ']' 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@518 -- # killprocess 4102015 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@954 -- # '[' -z 4102015 ']' 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@958 -- # kill -0 4102015 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@959 -- # uname 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4102015 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4102015' 00:32:58.328 killing process with pid 4102015 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@973 -- # kill 4102015 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@978 -- # wait 4102015 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-save 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:58.328 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-restore 00:32:58.589 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:58.589 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:58.589 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:58.589 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:58.589 19:29:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:00.503 00:33:00.503 real 0m24.557s 00:33:00.503 user 0m55.690s 00:33:00.503 sys 0m11.227s 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:33:00.503 ************************************ 00:33:00.503 END TEST nvmf_lvol 00:33:00.503 ************************************ 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:00.503 ************************************ 00:33:00.503 START TEST nvmf_lvs_grow 00:33:00.503 ************************************ 00:33:00.503 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:33:00.764 * Looking for test storage... 00:33:00.764 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1693 -- # lcov --version 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:00.764 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:33:00.765 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:00.765 --rc genhtml_branch_coverage=1 00:33:00.765 --rc genhtml_function_coverage=1 00:33:00.765 --rc genhtml_legend=1 00:33:00.765 --rc geninfo_all_blocks=1 00:33:00.765 --rc geninfo_unexecuted_blocks=1 00:33:00.765 00:33:00.765 ' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:33:00.765 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:00.765 --rc genhtml_branch_coverage=1 00:33:00.765 --rc genhtml_function_coverage=1 00:33:00.765 --rc genhtml_legend=1 00:33:00.765 --rc geninfo_all_blocks=1 00:33:00.765 --rc geninfo_unexecuted_blocks=1 00:33:00.765 00:33:00.765 ' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:33:00.765 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:00.765 --rc genhtml_branch_coverage=1 00:33:00.765 --rc genhtml_function_coverage=1 00:33:00.765 --rc genhtml_legend=1 00:33:00.765 --rc geninfo_all_blocks=1 00:33:00.765 --rc geninfo_unexecuted_blocks=1 00:33:00.765 00:33:00.765 ' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:33:00.765 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:00.765 --rc genhtml_branch_coverage=1 00:33:00.765 --rc genhtml_function_coverage=1 00:33:00.765 --rc genhtml_legend=1 00:33:00.765 --rc geninfo_all_blocks=1 00:33:00.765 --rc geninfo_unexecuted_blocks=1 00:33:00.765 00:33:00.765 ' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@476 -- # prepare_net_devs 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@438 -- # local -g is_hw=no 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # remove_spdk_ns 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:33:00.765 19:29:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:33:10.772 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:33:10.772 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:33:10.772 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:33:10.772 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:33:10.772 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:33:10.772 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:33:10.773 Found 0000:31:00.0 (0x8086 - 0x159b) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:33:10.773 Found 0000:31:00.1 (0x8086 - 0x159b) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:33:10.773 Found net devices under 0000:31:00.0: cvl_0_0 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:33:10.773 Found net devices under 0000:31:00.1: cvl_0_1 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # is_hw=yes 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:33:10.773 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:33:10.773 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.639 ms 00:33:10.773 00:33:10.773 --- 10.0.0.2 ping statistics --- 00:33:10.773 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:10.773 rtt min/avg/max/mdev = 0.639/0.639/0.639/0.000 ms 00:33:10.773 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:33:10.773 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:33:10.773 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.278 ms 00:33:10.773 00:33:10.773 --- 10.0.0.1 ping statistics --- 00:33:10.774 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:10.774 rtt min/avg/max/mdev = 0.278/0.278/0.278/0.000 ms 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@450 -- # return 0 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@726 -- # xtrace_disable 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@509 -- # nvmfpid=4109409 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@510 -- # waitforlisten 4109409 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # '[' -z 4109409 ']' 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:10.774 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:10.774 19:29:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:33:10.774 [2024-11-26 19:29:21.995921] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:33:10.774 [2024-11-26 19:29:21.996957] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:10.774 [2024-11-26 19:29:21.996997] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:10.774 [2024-11-26 19:29:22.085624] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:10.774 [2024-11-26 19:29:22.122490] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:10.774 [2024-11-26 19:29:22.122527] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:10.774 [2024-11-26 19:29:22.122535] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:10.774 [2024-11-26 19:29:22.122542] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:10.774 [2024-11-26 19:29:22.122549] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:10.774 [2024-11-26 19:29:22.123116] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:10.774 [2024-11-26 19:29:22.178813] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:33:10.774 [2024-11-26 19:29:22.179076] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:33:10.774 19:29:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:10.774 19:29:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@868 -- # return 0 00:33:10.774 19:29:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:33:10.774 19:29:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@732 -- # xtrace_disable 00:33:10.774 19:29:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:33:10.774 19:29:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:10.774 19:29:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:33:10.774 [2024-11-26 19:29:22.999559] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:33:10.774 ************************************ 00:33:10.774 START TEST lvs_grow_clean 00:33:10.774 ************************************ 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1129 -- # lvs_grow 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:33:10.774 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:33:11.034 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=336bb1e6-118e-430d-9743-9c45f549ef08 00:33:11.035 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:11.035 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:33:11.035 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:33:11.035 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:33:11.035 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 336bb1e6-118e-430d-9743-9c45f549ef08 lvol 150 00:33:11.296 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=898c6432-4f61-45a8-a393-d69ff35df165 00:33:11.296 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:11.296 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:33:11.556 [2024-11-26 19:29:23.943477] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:33:11.557 [2024-11-26 19:29:23.943546] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:33:11.557 true 00:33:11.557 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:11.557 19:29:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:33:11.557 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:33:11.557 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:33:11.817 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 898c6432-4f61-45a8-a393-d69ff35df165 00:33:12.079 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:33:12.079 [2024-11-26 19:29:24.591724] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:12.079 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=4109972 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 4109972 /var/tmp/bdevperf.sock 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # '[' -z 4109972 ']' 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:33:12.340 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:33:12.340 19:29:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:33:12.340 [2024-11-26 19:29:24.811878] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:12.340 [2024-11-26 19:29:24.811935] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4109972 ] 00:33:12.340 [2024-11-26 19:29:24.907231] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:12.340 [2024-11-26 19:29:24.944779] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:33:13.282 19:29:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:13.282 19:29:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@868 -- # return 0 00:33:13.282 19:29:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:33:13.543 Nvme0n1 00:33:13.543 19:29:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:33:13.543 [ 00:33:13.543 { 00:33:13.543 "name": "Nvme0n1", 00:33:13.543 "aliases": [ 00:33:13.543 "898c6432-4f61-45a8-a393-d69ff35df165" 00:33:13.543 ], 00:33:13.543 "product_name": "NVMe disk", 00:33:13.543 "block_size": 4096, 00:33:13.543 "num_blocks": 38912, 00:33:13.543 "uuid": "898c6432-4f61-45a8-a393-d69ff35df165", 00:33:13.543 "numa_id": 0, 00:33:13.543 "assigned_rate_limits": { 00:33:13.543 "rw_ios_per_sec": 0, 00:33:13.543 "rw_mbytes_per_sec": 0, 00:33:13.543 "r_mbytes_per_sec": 0, 00:33:13.543 "w_mbytes_per_sec": 0 00:33:13.543 }, 00:33:13.543 "claimed": false, 00:33:13.543 "zoned": false, 00:33:13.543 "supported_io_types": { 00:33:13.543 "read": true, 00:33:13.543 "write": true, 00:33:13.543 "unmap": true, 00:33:13.543 "flush": true, 00:33:13.543 "reset": true, 00:33:13.543 "nvme_admin": true, 00:33:13.543 "nvme_io": true, 00:33:13.543 "nvme_io_md": false, 00:33:13.543 "write_zeroes": true, 00:33:13.543 "zcopy": false, 00:33:13.543 "get_zone_info": false, 00:33:13.543 "zone_management": false, 00:33:13.543 "zone_append": false, 00:33:13.543 "compare": true, 00:33:13.543 "compare_and_write": true, 00:33:13.543 "abort": true, 00:33:13.543 "seek_hole": false, 00:33:13.543 "seek_data": false, 00:33:13.543 "copy": true, 00:33:13.543 "nvme_iov_md": false 00:33:13.543 }, 00:33:13.543 "memory_domains": [ 00:33:13.543 { 00:33:13.543 "dma_device_id": "system", 00:33:13.543 "dma_device_type": 1 00:33:13.543 } 00:33:13.543 ], 00:33:13.543 "driver_specific": { 00:33:13.543 "nvme": [ 00:33:13.543 { 00:33:13.543 "trid": { 00:33:13.543 "trtype": "TCP", 00:33:13.543 "adrfam": "IPv4", 00:33:13.543 "traddr": "10.0.0.2", 00:33:13.543 "trsvcid": "4420", 00:33:13.543 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:33:13.543 }, 00:33:13.543 "ctrlr_data": { 00:33:13.543 "cntlid": 1, 00:33:13.543 "vendor_id": "0x8086", 00:33:13.543 "model_number": "SPDK bdev Controller", 00:33:13.543 "serial_number": "SPDK0", 00:33:13.543 "firmware_revision": "25.01", 00:33:13.543 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:33:13.543 "oacs": { 00:33:13.543 "security": 0, 00:33:13.543 "format": 0, 00:33:13.543 "firmware": 0, 00:33:13.543 "ns_manage": 0 00:33:13.543 }, 00:33:13.543 "multi_ctrlr": true, 00:33:13.543 "ana_reporting": false 00:33:13.543 }, 00:33:13.543 "vs": { 00:33:13.543 "nvme_version": "1.3" 00:33:13.543 }, 00:33:13.543 "ns_data": { 00:33:13.543 "id": 1, 00:33:13.543 "can_share": true 00:33:13.543 } 00:33:13.543 } 00:33:13.543 ], 00:33:13.543 "mp_policy": "active_passive" 00:33:13.543 } 00:33:13.543 } 00:33:13.543 ] 00:33:13.543 19:29:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=4110146 00:33:13.543 19:29:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:33:13.543 19:29:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:33:13.804 Running I/O for 10 seconds... 00:33:14.746 Latency(us) 00:33:14.746 [2024-11-26T18:29:27.371Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:14.746 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:14.746 Nvme0n1 : 1.00 17790.00 69.49 0.00 0.00 0.00 0.00 0.00 00:33:14.746 [2024-11-26T18:29:27.371Z] =================================================================================================================== 00:33:14.746 [2024-11-26T18:29:27.371Z] Total : 17790.00 69.49 0.00 0.00 0.00 0.00 0.00 00:33:14.746 00:33:15.686 19:29:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:15.686 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:15.686 Nvme0n1 : 2.00 17848.50 69.72 0.00 0.00 0.00 0.00 0.00 00:33:15.686 [2024-11-26T18:29:28.311Z] =================================================================================================================== 00:33:15.686 [2024-11-26T18:29:28.311Z] Total : 17848.50 69.72 0.00 0.00 0.00 0.00 0.00 00:33:15.686 00:33:15.686 true 00:33:15.686 19:29:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:15.686 19:29:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:33:15.946 19:29:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:33:15.946 19:29:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:33:15.946 19:29:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 4110146 00:33:16.886 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:16.886 Nvme0n1 : 3.00 17868.00 69.80 0.00 0.00 0.00 0.00 0.00 00:33:16.886 [2024-11-26T18:29:29.511Z] =================================================================================================================== 00:33:16.886 [2024-11-26T18:29:29.511Z] Total : 17868.00 69.80 0.00 0.00 0.00 0.00 0.00 00:33:16.886 00:33:17.827 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:17.827 Nvme0n1 : 4.00 17909.50 69.96 0.00 0.00 0.00 0.00 0.00 00:33:17.827 [2024-11-26T18:29:30.452Z] =================================================================================================================== 00:33:17.827 [2024-11-26T18:29:30.452Z] Total : 17909.50 69.96 0.00 0.00 0.00 0.00 0.00 00:33:17.827 00:33:18.771 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:18.771 Nvme0n1 : 5.00 17925.20 70.02 0.00 0.00 0.00 0.00 0.00 00:33:18.771 [2024-11-26T18:29:31.396Z] =================================================================================================================== 00:33:18.771 [2024-11-26T18:29:31.396Z] Total : 17925.20 70.02 0.00 0.00 0.00 0.00 0.00 00:33:18.771 00:33:19.713 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:19.713 Nvme0n1 : 6.00 17940.50 70.08 0.00 0.00 0.00 0.00 0.00 00:33:19.713 [2024-11-26T18:29:32.338Z] =================================================================================================================== 00:33:19.713 [2024-11-26T18:29:32.338Z] Total : 17940.50 70.08 0.00 0.00 0.00 0.00 0.00 00:33:19.713 00:33:20.656 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:20.656 Nvme0n1 : 7.00 17962.86 70.17 0.00 0.00 0.00 0.00 0.00 00:33:20.656 [2024-11-26T18:29:33.281Z] =================================================================================================================== 00:33:20.656 [2024-11-26T18:29:33.281Z] Total : 17962.86 70.17 0.00 0.00 0.00 0.00 0.00 00:33:20.656 00:33:21.597 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:21.597 Nvme0n1 : 8.00 17971.75 70.20 0.00 0.00 0.00 0.00 0.00 00:33:21.597 [2024-11-26T18:29:34.222Z] =================================================================================================================== 00:33:21.597 [2024-11-26T18:29:34.222Z] Total : 17971.75 70.20 0.00 0.00 0.00 0.00 0.00 00:33:21.597 00:33:22.982 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:22.983 Nvme0n1 : 9.00 17978.67 70.23 0.00 0.00 0.00 0.00 0.00 00:33:22.983 [2024-11-26T18:29:35.608Z] =================================================================================================================== 00:33:22.983 [2024-11-26T18:29:35.608Z] Total : 17978.67 70.23 0.00 0.00 0.00 0.00 0.00 00:33:22.983 00:33:23.924 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:23.924 Nvme0n1 : 10.00 17984.20 70.25 0.00 0.00 0.00 0.00 0.00 00:33:23.924 [2024-11-26T18:29:36.549Z] =================================================================================================================== 00:33:23.924 [2024-11-26T18:29:36.549Z] Total : 17984.20 70.25 0.00 0.00 0.00 0.00 0.00 00:33:23.924 00:33:23.924 00:33:23.924 Latency(us) 00:33:23.924 [2024-11-26T18:29:36.549Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:23.924 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:23.924 Nvme0n1 : 10.00 17991.52 70.28 0.00 0.00 7110.82 2170.88 13271.04 00:33:23.924 [2024-11-26T18:29:36.549Z] =================================================================================================================== 00:33:23.924 [2024-11-26T18:29:36.549Z] Total : 17991.52 70.28 0.00 0.00 7110.82 2170.88 13271.04 00:33:23.924 { 00:33:23.924 "results": [ 00:33:23.924 { 00:33:23.924 "job": "Nvme0n1", 00:33:23.924 "core_mask": "0x2", 00:33:23.924 "workload": "randwrite", 00:33:23.924 "status": "finished", 00:33:23.924 "queue_depth": 128, 00:33:23.924 "io_size": 4096, 00:33:23.924 "runtime": 10.003047, 00:33:23.924 "iops": 17991.51798447013, 00:33:23.924 "mibps": 70.27936712683645, 00:33:23.924 "io_failed": 0, 00:33:23.924 "io_timeout": 0, 00:33:23.924 "avg_latency_us": 7110.820853549665, 00:33:23.924 "min_latency_us": 2170.88, 00:33:23.924 "max_latency_us": 13271.04 00:33:23.924 } 00:33:23.924 ], 00:33:23.924 "core_count": 1 00:33:23.924 } 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 4109972 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # '[' -z 4109972 ']' 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@958 -- # kill -0 4109972 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # uname 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4109972 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4109972' 00:33:23.924 killing process with pid 4109972 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@973 -- # kill 4109972 00:33:23.924 Received shutdown signal, test time was about 10.000000 seconds 00:33:23.924 00:33:23.924 Latency(us) 00:33:23.924 [2024-11-26T18:29:36.549Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:23.924 [2024-11-26T18:29:36.549Z] =================================================================================================================== 00:33:23.924 [2024-11-26T18:29:36.549Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@978 -- # wait 4109972 00:33:23.924 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:24.185 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:33:24.185 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:24.185 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:33:24.445 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:33:24.445 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:33:24.445 19:29:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:33:24.445 [2024-11-26 19:29:37.039490] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # local es=0 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:24.706 request: 00:33:24.706 { 00:33:24.706 "uuid": "336bb1e6-118e-430d-9743-9c45f549ef08", 00:33:24.706 "method": "bdev_lvol_get_lvstores", 00:33:24.706 "req_id": 1 00:33:24.706 } 00:33:24.706 Got JSON-RPC error response 00:33:24.706 response: 00:33:24.706 { 00:33:24.706 "code": -19, 00:33:24.706 "message": "No such device" 00:33:24.706 } 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # es=1 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:33:24.706 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:33:24.967 aio_bdev 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 898c6432-4f61-45a8-a393-d69ff35df165 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@903 -- # local bdev_name=898c6432-4f61-45a8-a393-d69ff35df165 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@905 -- # local i 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:33:24.967 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 898c6432-4f61-45a8-a393-d69ff35df165 -t 2000 00:33:25.228 [ 00:33:25.228 { 00:33:25.228 "name": "898c6432-4f61-45a8-a393-d69ff35df165", 00:33:25.228 "aliases": [ 00:33:25.228 "lvs/lvol" 00:33:25.228 ], 00:33:25.228 "product_name": "Logical Volume", 00:33:25.228 "block_size": 4096, 00:33:25.228 "num_blocks": 38912, 00:33:25.228 "uuid": "898c6432-4f61-45a8-a393-d69ff35df165", 00:33:25.228 "assigned_rate_limits": { 00:33:25.228 "rw_ios_per_sec": 0, 00:33:25.228 "rw_mbytes_per_sec": 0, 00:33:25.228 "r_mbytes_per_sec": 0, 00:33:25.228 "w_mbytes_per_sec": 0 00:33:25.228 }, 00:33:25.228 "claimed": false, 00:33:25.228 "zoned": false, 00:33:25.228 "supported_io_types": { 00:33:25.228 "read": true, 00:33:25.228 "write": true, 00:33:25.228 "unmap": true, 00:33:25.228 "flush": false, 00:33:25.228 "reset": true, 00:33:25.228 "nvme_admin": false, 00:33:25.228 "nvme_io": false, 00:33:25.229 "nvme_io_md": false, 00:33:25.229 "write_zeroes": true, 00:33:25.229 "zcopy": false, 00:33:25.229 "get_zone_info": false, 00:33:25.229 "zone_management": false, 00:33:25.229 "zone_append": false, 00:33:25.229 "compare": false, 00:33:25.229 "compare_and_write": false, 00:33:25.229 "abort": false, 00:33:25.229 "seek_hole": true, 00:33:25.229 "seek_data": true, 00:33:25.229 "copy": false, 00:33:25.229 "nvme_iov_md": false 00:33:25.229 }, 00:33:25.229 "driver_specific": { 00:33:25.229 "lvol": { 00:33:25.229 "lvol_store_uuid": "336bb1e6-118e-430d-9743-9c45f549ef08", 00:33:25.229 "base_bdev": "aio_bdev", 00:33:25.229 "thin_provision": false, 00:33:25.229 "num_allocated_clusters": 38, 00:33:25.229 "snapshot": false, 00:33:25.229 "clone": false, 00:33:25.229 "esnap_clone": false 00:33:25.229 } 00:33:25.229 } 00:33:25.229 } 00:33:25.229 ] 00:33:25.229 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@911 -- # return 0 00:33:25.229 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:25.229 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:33:25.489 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:33:25.489 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:25.489 19:29:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:33:25.489 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:33:25.489 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 898c6432-4f61-45a8-a393-d69ff35df165 00:33:25.750 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 336bb1e6-118e-430d-9743-9c45f549ef08 00:33:26.011 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:33:26.011 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:26.011 00:33:26.011 real 0m15.536s 00:33:26.011 user 0m15.190s 00:33:26.011 sys 0m1.307s 00:33:26.011 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:26.011 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:33:26.012 ************************************ 00:33:26.012 END TEST lvs_grow_clean 00:33:26.012 ************************************ 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:33:26.272 ************************************ 00:33:26.272 START TEST lvs_grow_dirty 00:33:26.272 ************************************ 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1129 -- # lvs_grow dirty 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:33:26.272 19:29:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:33:26.532 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=37841dc1-f069-4aa4-be61-e27e615011b2 00:33:26.532 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:26.532 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:33:26.792 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:33:26.792 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:33:26.792 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 37841dc1-f069-4aa4-be61-e27e615011b2 lvol 150 00:33:26.792 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 00:33:26.792 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:26.792 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:33:27.052 [2024-11-26 19:29:39.515445] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:33:27.052 [2024-11-26 19:29:39.515514] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:33:27.052 true 00:33:27.052 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:27.052 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:33:27.313 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:33:27.313 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:33:27.313 19:29:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 00:33:27.574 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:33:27.574 [2024-11-26 19:29:40.188044] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=4112876 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 4112876 /var/tmp/bdevperf.sock 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 4112876 ']' 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:33:27.835 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:27.835 19:29:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:33:27.835 [2024-11-26 19:29:40.440227] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:27.835 [2024-11-26 19:29:40.440284] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4112876 ] 00:33:28.095 [2024-11-26 19:29:40.533783] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:28.095 [2024-11-26 19:29:40.564132] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:33:28.667 19:29:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:28.667 19:29:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:33:28.667 19:29:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:33:29.239 Nvme0n1 00:33:29.239 19:29:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:33:29.239 [ 00:33:29.239 { 00:33:29.239 "name": "Nvme0n1", 00:33:29.239 "aliases": [ 00:33:29.239 "8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250" 00:33:29.239 ], 00:33:29.239 "product_name": "NVMe disk", 00:33:29.239 "block_size": 4096, 00:33:29.239 "num_blocks": 38912, 00:33:29.239 "uuid": "8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250", 00:33:29.239 "numa_id": 0, 00:33:29.239 "assigned_rate_limits": { 00:33:29.239 "rw_ios_per_sec": 0, 00:33:29.239 "rw_mbytes_per_sec": 0, 00:33:29.239 "r_mbytes_per_sec": 0, 00:33:29.239 "w_mbytes_per_sec": 0 00:33:29.239 }, 00:33:29.239 "claimed": false, 00:33:29.239 "zoned": false, 00:33:29.239 "supported_io_types": { 00:33:29.239 "read": true, 00:33:29.239 "write": true, 00:33:29.239 "unmap": true, 00:33:29.239 "flush": true, 00:33:29.239 "reset": true, 00:33:29.239 "nvme_admin": true, 00:33:29.239 "nvme_io": true, 00:33:29.239 "nvme_io_md": false, 00:33:29.239 "write_zeroes": true, 00:33:29.239 "zcopy": false, 00:33:29.239 "get_zone_info": false, 00:33:29.239 "zone_management": false, 00:33:29.239 "zone_append": false, 00:33:29.239 "compare": true, 00:33:29.239 "compare_and_write": true, 00:33:29.239 "abort": true, 00:33:29.239 "seek_hole": false, 00:33:29.239 "seek_data": false, 00:33:29.239 "copy": true, 00:33:29.239 "nvme_iov_md": false 00:33:29.239 }, 00:33:29.239 "memory_domains": [ 00:33:29.239 { 00:33:29.239 "dma_device_id": "system", 00:33:29.239 "dma_device_type": 1 00:33:29.239 } 00:33:29.239 ], 00:33:29.239 "driver_specific": { 00:33:29.239 "nvme": [ 00:33:29.239 { 00:33:29.239 "trid": { 00:33:29.239 "trtype": "TCP", 00:33:29.239 "adrfam": "IPv4", 00:33:29.239 "traddr": "10.0.0.2", 00:33:29.239 "trsvcid": "4420", 00:33:29.239 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:33:29.239 }, 00:33:29.239 "ctrlr_data": { 00:33:29.239 "cntlid": 1, 00:33:29.239 "vendor_id": "0x8086", 00:33:29.239 "model_number": "SPDK bdev Controller", 00:33:29.239 "serial_number": "SPDK0", 00:33:29.239 "firmware_revision": "25.01", 00:33:29.239 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:33:29.239 "oacs": { 00:33:29.239 "security": 0, 00:33:29.239 "format": 0, 00:33:29.239 "firmware": 0, 00:33:29.239 "ns_manage": 0 00:33:29.239 }, 00:33:29.239 "multi_ctrlr": true, 00:33:29.239 "ana_reporting": false 00:33:29.239 }, 00:33:29.239 "vs": { 00:33:29.239 "nvme_version": "1.3" 00:33:29.239 }, 00:33:29.239 "ns_data": { 00:33:29.239 "id": 1, 00:33:29.239 "can_share": true 00:33:29.239 } 00:33:29.239 } 00:33:29.239 ], 00:33:29.239 "mp_policy": "active_passive" 00:33:29.239 } 00:33:29.239 } 00:33:29.239 ] 00:33:29.239 19:29:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:33:29.239 19:29:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=4113211 00:33:29.239 19:29:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:33:29.239 Running I/O for 10 seconds... 00:33:30.622 Latency(us) 00:33:30.622 [2024-11-26T18:29:43.247Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:30.622 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:30.622 Nvme0n1 : 1.00 17720.00 69.22 0.00 0.00 0.00 0.00 0.00 00:33:30.622 [2024-11-26T18:29:43.247Z] =================================================================================================================== 00:33:30.622 [2024-11-26T18:29:43.247Z] Total : 17720.00 69.22 0.00 0.00 0.00 0.00 0.00 00:33:30.622 00:33:31.192 19:29:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:31.452 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:31.452 Nvme0n1 : 2.00 17822.00 69.62 0.00 0.00 0.00 0.00 0.00 00:33:31.452 [2024-11-26T18:29:44.077Z] =================================================================================================================== 00:33:31.452 [2024-11-26T18:29:44.077Z] Total : 17822.00 69.62 0.00 0.00 0.00 0.00 0.00 00:33:31.452 00:33:31.452 true 00:33:31.452 19:29:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:31.452 19:29:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:33:31.713 19:29:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:33:31.713 19:29:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:33:31.713 19:29:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 4113211 00:33:32.283 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:32.283 Nvme0n1 : 3.00 17865.67 69.79 0.00 0.00 0.00 0.00 0.00 00:33:32.283 [2024-11-26T18:29:44.908Z] =================================================================================================================== 00:33:32.283 [2024-11-26T18:29:44.908Z] Total : 17865.67 69.79 0.00 0.00 0.00 0.00 0.00 00:33:32.283 00:33:33.663 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:33.663 Nvme0n1 : 4.00 17876.00 69.83 0.00 0.00 0.00 0.00 0.00 00:33:33.663 [2024-11-26T18:29:46.288Z] =================================================================================================================== 00:33:33.663 [2024-11-26T18:29:46.288Z] Total : 17876.00 69.83 0.00 0.00 0.00 0.00 0.00 00:33:33.663 00:33:34.602 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:34.602 Nvme0n1 : 5.00 17907.60 69.95 0.00 0.00 0.00 0.00 0.00 00:33:34.602 [2024-11-26T18:29:47.227Z] =================================================================================================================== 00:33:34.602 [2024-11-26T18:29:47.227Z] Total : 17907.60 69.95 0.00 0.00 0.00 0.00 0.00 00:33:34.602 00:33:35.539 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:35.539 Nvme0n1 : 6.00 17928.67 70.03 0.00 0.00 0.00 0.00 0.00 00:33:35.540 [2024-11-26T18:29:48.165Z] =================================================================================================================== 00:33:35.540 [2024-11-26T18:29:48.165Z] Total : 17928.67 70.03 0.00 0.00 0.00 0.00 0.00 00:33:35.540 00:33:36.478 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:36.478 Nvme0n1 : 7.00 17943.71 70.09 0.00 0.00 0.00 0.00 0.00 00:33:36.478 [2024-11-26T18:29:49.104Z] =================================================================================================================== 00:33:36.479 [2024-11-26T18:29:49.104Z] Total : 17943.71 70.09 0.00 0.00 0.00 0.00 0.00 00:33:36.479 00:33:37.418 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:37.418 Nvme0n1 : 8.00 17955.00 70.14 0.00 0.00 0.00 0.00 0.00 00:33:37.418 [2024-11-26T18:29:50.043Z] =================================================================================================================== 00:33:37.418 [2024-11-26T18:29:50.043Z] Total : 17955.00 70.14 0.00 0.00 0.00 0.00 0.00 00:33:37.418 00:33:38.361 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:38.361 Nvme0n1 : 9.00 17963.78 70.17 0.00 0.00 0.00 0.00 0.00 00:33:38.361 [2024-11-26T18:29:50.986Z] =================================================================================================================== 00:33:38.361 [2024-11-26T18:29:50.986Z] Total : 17963.78 70.17 0.00 0.00 0.00 0.00 0.00 00:33:38.361 00:33:39.302 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:39.302 Nvme0n1 : 10.00 17970.80 70.20 0.00 0.00 0.00 0.00 0.00 00:33:39.302 [2024-11-26T18:29:51.927Z] =================================================================================================================== 00:33:39.302 [2024-11-26T18:29:51.927Z] Total : 17970.80 70.20 0.00 0.00 0.00 0.00 0.00 00:33:39.302 00:33:39.302 00:33:39.302 Latency(us) 00:33:39.302 [2024-11-26T18:29:51.927Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:39.302 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:33:39.302 Nvme0n1 : 10.00 17978.54 70.23 0.00 0.00 7116.33 1665.71 13926.40 00:33:39.302 [2024-11-26T18:29:51.927Z] =================================================================================================================== 00:33:39.302 [2024-11-26T18:29:51.927Z] Total : 17978.54 70.23 0.00 0.00 7116.33 1665.71 13926.40 00:33:39.302 { 00:33:39.302 "results": [ 00:33:39.302 { 00:33:39.302 "job": "Nvme0n1", 00:33:39.302 "core_mask": "0x2", 00:33:39.302 "workload": "randwrite", 00:33:39.302 "status": "finished", 00:33:39.302 "queue_depth": 128, 00:33:39.302 "io_size": 4096, 00:33:39.302 "runtime": 10.002812, 00:33:39.302 "iops": 17978.544433305353, 00:33:39.302 "mibps": 70.22868919259903, 00:33:39.302 "io_failed": 0, 00:33:39.302 "io_timeout": 0, 00:33:39.302 "avg_latency_us": 7116.327714287833, 00:33:39.302 "min_latency_us": 1665.7066666666667, 00:33:39.302 "max_latency_us": 13926.4 00:33:39.302 } 00:33:39.302 ], 00:33:39.302 "core_count": 1 00:33:39.302 } 00:33:39.302 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 4112876 00:33:39.302 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # '[' -z 4112876 ']' 00:33:39.302 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@958 -- # kill -0 4112876 00:33:39.302 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # uname 00:33:39.302 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:39.302 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4112876 00:33:39.562 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:33:39.562 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:33:39.562 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4112876' 00:33:39.562 killing process with pid 4112876 00:33:39.562 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@973 -- # kill 4112876 00:33:39.562 Received shutdown signal, test time was about 10.000000 seconds 00:33:39.562 00:33:39.562 Latency(us) 00:33:39.562 [2024-11-26T18:29:52.187Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:39.562 [2024-11-26T18:29:52.187Z] =================================================================================================================== 00:33:39.562 [2024-11-26T18:29:52.187Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:33:39.562 19:29:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@978 -- # wait 4112876 00:33:39.562 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:39.822 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:33:39.822 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:39.822 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 4109409 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 4109409 00:33:40.082 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 4109409 Killed "${NVMF_APP[@]}" "$@" 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@726 -- # xtrace_disable 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # nvmfpid=4115229 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # waitforlisten 4115229 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 4115229 ']' 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:40.082 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:40.082 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:33:40.082 [2024-11-26 19:29:52.649232] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:33:40.082 [2024-11-26 19:29:52.649956] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:40.082 [2024-11-26 19:29:52.649992] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:40.343 [2024-11-26 19:29:52.723224] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:40.343 [2024-11-26 19:29:52.757487] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:40.344 [2024-11-26 19:29:52.757521] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:40.344 [2024-11-26 19:29:52.757529] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:40.344 [2024-11-26 19:29:52.757536] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:40.344 [2024-11-26 19:29:52.757542] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:40.344 [2024-11-26 19:29:52.758099] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:40.344 [2024-11-26 19:29:52.813565] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:33:40.344 [2024-11-26 19:29:52.813821] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:33:40.344 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:40.344 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:33:40.344 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:33:40.344 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@732 -- # xtrace_disable 00:33:40.344 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:33:40.344 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:40.344 19:29:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:33:40.605 [2024-11-26 19:29:53.032969] blobstore.c:4896:bs_recover: *NOTICE*: Performing recovery on blobstore 00:33:40.605 [2024-11-26 19:29:53.033063] blobstore.c:4843:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:33:40.605 [2024-11-26 19:29:53.033093] blobstore.c:4843:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev 8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:33:40.605 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 -t 2000 00:33:40.865 [ 00:33:40.865 { 00:33:40.865 "name": "8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250", 00:33:40.865 "aliases": [ 00:33:40.865 "lvs/lvol" 00:33:40.865 ], 00:33:40.865 "product_name": "Logical Volume", 00:33:40.865 "block_size": 4096, 00:33:40.865 "num_blocks": 38912, 00:33:40.865 "uuid": "8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250", 00:33:40.865 "assigned_rate_limits": { 00:33:40.865 "rw_ios_per_sec": 0, 00:33:40.865 "rw_mbytes_per_sec": 0, 00:33:40.865 "r_mbytes_per_sec": 0, 00:33:40.865 "w_mbytes_per_sec": 0 00:33:40.865 }, 00:33:40.865 "claimed": false, 00:33:40.865 "zoned": false, 00:33:40.865 "supported_io_types": { 00:33:40.865 "read": true, 00:33:40.865 "write": true, 00:33:40.865 "unmap": true, 00:33:40.865 "flush": false, 00:33:40.865 "reset": true, 00:33:40.865 "nvme_admin": false, 00:33:40.865 "nvme_io": false, 00:33:40.865 "nvme_io_md": false, 00:33:40.865 "write_zeroes": true, 00:33:40.865 "zcopy": false, 00:33:40.865 "get_zone_info": false, 00:33:40.865 "zone_management": false, 00:33:40.865 "zone_append": false, 00:33:40.865 "compare": false, 00:33:40.865 "compare_and_write": false, 00:33:40.865 "abort": false, 00:33:40.865 "seek_hole": true, 00:33:40.865 "seek_data": true, 00:33:40.865 "copy": false, 00:33:40.865 "nvme_iov_md": false 00:33:40.865 }, 00:33:40.865 "driver_specific": { 00:33:40.865 "lvol": { 00:33:40.865 "lvol_store_uuid": "37841dc1-f069-4aa4-be61-e27e615011b2", 00:33:40.865 "base_bdev": "aio_bdev", 00:33:40.865 "thin_provision": false, 00:33:40.865 "num_allocated_clusters": 38, 00:33:40.865 "snapshot": false, 00:33:40.865 "clone": false, 00:33:40.865 "esnap_clone": false 00:33:40.865 } 00:33:40.865 } 00:33:40.865 } 00:33:40.865 ] 00:33:40.865 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:33:40.865 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:40.865 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:33:41.126 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:33:41.126 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:41.126 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:33:41.126 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:33:41.126 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:33:41.385 [2024-11-26 19:29:53.854623] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # local es=0 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:33:41.385 19:29:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:41.645 request: 00:33:41.645 { 00:33:41.645 "uuid": "37841dc1-f069-4aa4-be61-e27e615011b2", 00:33:41.645 "method": "bdev_lvol_get_lvstores", 00:33:41.645 "req_id": 1 00:33:41.645 } 00:33:41.645 Got JSON-RPC error response 00:33:41.646 response: 00:33:41.646 { 00:33:41.646 "code": -19, 00:33:41.646 "message": "No such device" 00:33:41.646 } 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # es=1 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:33:41.646 aio_bdev 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:41.646 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:33:41.906 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 -t 2000 00:33:41.906 [ 00:33:41.906 { 00:33:41.906 "name": "8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250", 00:33:41.906 "aliases": [ 00:33:41.906 "lvs/lvol" 00:33:41.906 ], 00:33:41.906 "product_name": "Logical Volume", 00:33:41.906 "block_size": 4096, 00:33:41.906 "num_blocks": 38912, 00:33:41.906 "uuid": "8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250", 00:33:41.906 "assigned_rate_limits": { 00:33:41.906 "rw_ios_per_sec": 0, 00:33:41.906 "rw_mbytes_per_sec": 0, 00:33:41.907 "r_mbytes_per_sec": 0, 00:33:41.907 "w_mbytes_per_sec": 0 00:33:41.907 }, 00:33:41.907 "claimed": false, 00:33:41.907 "zoned": false, 00:33:41.907 "supported_io_types": { 00:33:41.907 "read": true, 00:33:41.907 "write": true, 00:33:41.907 "unmap": true, 00:33:41.907 "flush": false, 00:33:41.907 "reset": true, 00:33:41.907 "nvme_admin": false, 00:33:41.907 "nvme_io": false, 00:33:41.907 "nvme_io_md": false, 00:33:41.907 "write_zeroes": true, 00:33:41.907 "zcopy": false, 00:33:41.907 "get_zone_info": false, 00:33:41.907 "zone_management": false, 00:33:41.907 "zone_append": false, 00:33:41.907 "compare": false, 00:33:41.907 "compare_and_write": false, 00:33:41.907 "abort": false, 00:33:41.907 "seek_hole": true, 00:33:41.907 "seek_data": true, 00:33:41.907 "copy": false, 00:33:41.907 "nvme_iov_md": false 00:33:41.907 }, 00:33:41.907 "driver_specific": { 00:33:41.907 "lvol": { 00:33:41.907 "lvol_store_uuid": "37841dc1-f069-4aa4-be61-e27e615011b2", 00:33:41.907 "base_bdev": "aio_bdev", 00:33:41.907 "thin_provision": false, 00:33:41.907 "num_allocated_clusters": 38, 00:33:41.907 "snapshot": false, 00:33:41.907 "clone": false, 00:33:41.907 "esnap_clone": false 00:33:41.907 } 00:33:41.907 } 00:33:41.907 } 00:33:41.907 ] 00:33:41.907 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:33:42.167 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:33:42.167 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:42.167 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:33:42.167 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:42.167 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:33:42.427 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:33:42.427 19:29:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 8ff3bbcb-80fb-489b-85b7-ff2e7c6f8250 00:33:42.427 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 37841dc1-f069-4aa4-be61-e27e615011b2 00:33:42.688 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:33:42.948 00:33:42.948 real 0m16.690s 00:33:42.948 user 0m35.163s 00:33:42.948 sys 0m2.789s 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:33:42.948 ************************************ 00:33:42.948 END TEST lvs_grow_dirty 00:33:42.948 ************************************ 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@812 -- # type=--id 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@813 -- # id=0 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@824 -- # for n in $shm_files 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:33:42.948 nvmf_trace.0 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@827 -- # return 0 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@516 -- # nvmfcleanup 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:33:42.948 rmmod nvme_tcp 00:33:42.948 rmmod nvme_fabrics 00:33:42.948 rmmod nvme_keyring 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@517 -- # '[' -n 4115229 ']' 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@518 -- # killprocess 4115229 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # '[' -z 4115229 ']' 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@958 -- # kill -0 4115229 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # uname 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:42.948 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4115229 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4115229' 00:33:43.208 killing process with pid 4115229 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@973 -- # kill 4115229 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@978 -- # wait 4115229 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-save 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-restore 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:43.208 19:29:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:45.750 00:33:45.750 real 0m44.696s 00:33:45.750 user 0m53.577s 00:33:45.750 sys 0m10.990s 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:33:45.750 ************************************ 00:33:45.750 END TEST nvmf_lvs_grow 00:33:45.750 ************************************ 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:45.750 ************************************ 00:33:45.750 START TEST nvmf_bdev_io_wait 00:33:45.750 ************************************ 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:33:45.750 * Looking for test storage... 00:33:45.750 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1693 -- # lcov --version 00:33:45.750 19:29:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:33:45.750 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:45.750 --rc genhtml_branch_coverage=1 00:33:45.750 --rc genhtml_function_coverage=1 00:33:45.750 --rc genhtml_legend=1 00:33:45.750 --rc geninfo_all_blocks=1 00:33:45.750 --rc geninfo_unexecuted_blocks=1 00:33:45.750 00:33:45.750 ' 00:33:45.750 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:33:45.750 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:45.750 --rc genhtml_branch_coverage=1 00:33:45.750 --rc genhtml_function_coverage=1 00:33:45.750 --rc genhtml_legend=1 00:33:45.750 --rc geninfo_all_blocks=1 00:33:45.750 --rc geninfo_unexecuted_blocks=1 00:33:45.750 00:33:45.750 ' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:33:45.751 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:45.751 --rc genhtml_branch_coverage=1 00:33:45.751 --rc genhtml_function_coverage=1 00:33:45.751 --rc genhtml_legend=1 00:33:45.751 --rc geninfo_all_blocks=1 00:33:45.751 --rc geninfo_unexecuted_blocks=1 00:33:45.751 00:33:45.751 ' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:33:45.751 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:45.751 --rc genhtml_branch_coverage=1 00:33:45.751 --rc genhtml_function_coverage=1 00:33:45.751 --rc genhtml_legend=1 00:33:45.751 --rc geninfo_all_blocks=1 00:33:45.751 --rc geninfo_unexecuted_blocks=1 00:33:45.751 00:33:45.751 ' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # prepare_net_devs 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # local -g is_hw=no 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # remove_spdk_ns 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:33:45.751 19:29:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:53.898 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:33:53.898 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:33:53.898 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:33:53.898 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:33:53.899 Found 0000:31:00.0 (0x8086 - 0x159b) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:33:53.899 Found 0000:31:00.1 (0x8086 - 0x159b) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:33:53.899 Found net devices under 0000:31:00.0: cvl_0_0 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:33:53.899 Found net devices under 0000:31:00.1: cvl_0_1 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # is_hw=yes 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:33:53.899 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:33:54.160 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:33:54.160 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.589 ms 00:33:54.160 00:33:54.160 --- 10.0.0.2 ping statistics --- 00:33:54.160 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:54.160 rtt min/avg/max/mdev = 0.589/0.589/0.589/0.000 ms 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:33:54.160 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:33:54.160 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.300 ms 00:33:54.160 00:33:54.160 --- 10.0.0.1 ping statistics --- 00:33:54.160 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:54.160 rtt min/avg/max/mdev = 0.300/0.300/0.300/0.000 ms 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@450 -- # return 0 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@726 -- # xtrace_disable 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # nvmfpid=4120735 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # waitforlisten 4120735 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF --wait-for-rpc 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # '[' -z 4120735 ']' 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:54.160 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:54.160 19:30:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:54.160 [2024-11-26 19:30:06.673867] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:33:54.160 [2024-11-26 19:30:06.674871] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:54.161 [2024-11-26 19:30:06.674908] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:54.161 [2024-11-26 19:30:06.759965] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:33:54.422 [2024-11-26 19:30:06.797962] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:54.422 [2024-11-26 19:30:06.797993] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:54.422 [2024-11-26 19:30:06.798001] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:54.422 [2024-11-26 19:30:06.798007] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:54.422 [2024-11-26 19:30:06.798014] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:54.422 [2024-11-26 19:30:06.799513] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:33:54.422 [2024-11-26 19:30:06.799628] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:33:54.422 [2024-11-26 19:30:06.799782] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:54.422 [2024-11-26 19:30:06.799783] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:33:54.422 [2024-11-26 19:30:06.800036] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@868 -- # return 0 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@732 -- # xtrace_disable 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:54.993 [2024-11-26 19:30:07.580412] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:33:54.993 [2024-11-26 19:30:07.580537] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:33:54.993 [2024-11-26 19:30:07.581308] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:33:54.993 [2024-11-26 19:30:07.581370] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:54.993 [2024-11-26 19:30:07.592536] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:54.993 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:55.260 Malloc0 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:55.261 [2024-11-26 19:30:07.656393] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=4120933 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=4120936 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:33:55.261 { 00:33:55.261 "params": { 00:33:55.261 "name": "Nvme$subsystem", 00:33:55.261 "trtype": "$TEST_TRANSPORT", 00:33:55.261 "traddr": "$NVMF_FIRST_TARGET_IP", 00:33:55.261 "adrfam": "ipv4", 00:33:55.261 "trsvcid": "$NVMF_PORT", 00:33:55.261 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:33:55.261 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:33:55.261 "hdgst": ${hdgst:-false}, 00:33:55.261 "ddgst": ${ddgst:-false} 00:33:55.261 }, 00:33:55.261 "method": "bdev_nvme_attach_controller" 00:33:55.261 } 00:33:55.261 EOF 00:33:55.261 )") 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=4120939 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:33:55.261 { 00:33:55.261 "params": { 00:33:55.261 "name": "Nvme$subsystem", 00:33:55.261 "trtype": "$TEST_TRANSPORT", 00:33:55.261 "traddr": "$NVMF_FIRST_TARGET_IP", 00:33:55.261 "adrfam": "ipv4", 00:33:55.261 "trsvcid": "$NVMF_PORT", 00:33:55.261 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:33:55.261 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:33:55.261 "hdgst": ${hdgst:-false}, 00:33:55.261 "ddgst": ${ddgst:-false} 00:33:55.261 }, 00:33:55.261 "method": "bdev_nvme_attach_controller" 00:33:55.261 } 00:33:55.261 EOF 00:33:55.261 )") 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=4120942 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:33:55.261 { 00:33:55.261 "params": { 00:33:55.261 "name": "Nvme$subsystem", 00:33:55.261 "trtype": "$TEST_TRANSPORT", 00:33:55.261 "traddr": "$NVMF_FIRST_TARGET_IP", 00:33:55.261 "adrfam": "ipv4", 00:33:55.261 "trsvcid": "$NVMF_PORT", 00:33:55.261 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:33:55.261 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:33:55.261 "hdgst": ${hdgst:-false}, 00:33:55.261 "ddgst": ${ddgst:-false} 00:33:55.261 }, 00:33:55.261 "method": "bdev_nvme_attach_controller" 00:33:55.261 } 00:33:55.261 EOF 00:33:55.261 )") 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:33:55.261 { 00:33:55.261 "params": { 00:33:55.261 "name": "Nvme$subsystem", 00:33:55.261 "trtype": "$TEST_TRANSPORT", 00:33:55.261 "traddr": "$NVMF_FIRST_TARGET_IP", 00:33:55.261 "adrfam": "ipv4", 00:33:55.261 "trsvcid": "$NVMF_PORT", 00:33:55.261 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:33:55.261 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:33:55.261 "hdgst": ${hdgst:-false}, 00:33:55.261 "ddgst": ${ddgst:-false} 00:33:55.261 }, 00:33:55.261 "method": "bdev_nvme_attach_controller" 00:33:55.261 } 00:33:55.261 EOF 00:33:55.261 )") 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 4120933 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:33:55.261 "params": { 00:33:55.261 "name": "Nvme1", 00:33:55.261 "trtype": "tcp", 00:33:55.261 "traddr": "10.0.0.2", 00:33:55.261 "adrfam": "ipv4", 00:33:55.261 "trsvcid": "4420", 00:33:55.261 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:33:55.261 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:33:55.261 "hdgst": false, 00:33:55.261 "ddgst": false 00:33:55.261 }, 00:33:55.261 "method": "bdev_nvme_attach_controller" 00:33:55.261 }' 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:33:55.261 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:33:55.261 "params": { 00:33:55.261 "name": "Nvme1", 00:33:55.261 "trtype": "tcp", 00:33:55.261 "traddr": "10.0.0.2", 00:33:55.261 "adrfam": "ipv4", 00:33:55.262 "trsvcid": "4420", 00:33:55.262 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:33:55.262 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:33:55.262 "hdgst": false, 00:33:55.262 "ddgst": false 00:33:55.262 }, 00:33:55.262 "method": "bdev_nvme_attach_controller" 00:33:55.262 }' 00:33:55.262 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:33:55.262 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:33:55.262 "params": { 00:33:55.262 "name": "Nvme1", 00:33:55.262 "trtype": "tcp", 00:33:55.262 "traddr": "10.0.0.2", 00:33:55.262 "adrfam": "ipv4", 00:33:55.262 "trsvcid": "4420", 00:33:55.262 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:33:55.262 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:33:55.262 "hdgst": false, 00:33:55.262 "ddgst": false 00:33:55.262 }, 00:33:55.262 "method": "bdev_nvme_attach_controller" 00:33:55.262 }' 00:33:55.262 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:33:55.262 19:30:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:33:55.262 "params": { 00:33:55.262 "name": "Nvme1", 00:33:55.262 "trtype": "tcp", 00:33:55.262 "traddr": "10.0.0.2", 00:33:55.262 "adrfam": "ipv4", 00:33:55.262 "trsvcid": "4420", 00:33:55.262 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:33:55.262 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:33:55.262 "hdgst": false, 00:33:55.262 "ddgst": false 00:33:55.262 }, 00:33:55.262 "method": "bdev_nvme_attach_controller" 00:33:55.262 }' 00:33:55.262 [2024-11-26 19:30:07.710747] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:55.262 [2024-11-26 19:30:07.710801] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:33:55.262 [2024-11-26 19:30:07.713977] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:55.262 [2024-11-26 19:30:07.714023] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:33:55.262 [2024-11-26 19:30:07.714650] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:55.262 [2024-11-26 19:30:07.714693] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:33:55.262 [2024-11-26 19:30:07.717065] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:33:55.262 [2024-11-26 19:30:07.717113] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:33:55.562 [2024-11-26 19:30:07.879306] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:55.562 [2024-11-26 19:30:07.908417] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:33:55.562 [2024-11-26 19:30:07.934825] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:55.562 [2024-11-26 19:30:07.963913] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:33:55.562 [2024-11-26 19:30:07.978995] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:55.562 [2024-11-26 19:30:08.007600] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:33:55.562 [2024-11-26 19:30:08.028561] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:55.562 [2024-11-26 19:30:08.056817] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:33:55.562 Running I/O for 1 seconds... 00:33:55.861 Running I/O for 1 seconds... 00:33:55.861 Running I/O for 1 seconds... 00:33:55.861 Running I/O for 1 seconds... 00:33:56.831 182960.00 IOPS, 714.69 MiB/s 00:33:56.831 Latency(us) 00:33:56.831 [2024-11-26T18:30:09.456Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:56.831 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:33:56.831 Nvme1n1 : 1.00 182601.70 713.29 0.00 0.00 697.35 298.67 1966.08 00:33:56.831 [2024-11-26T18:30:09.456Z] =================================================================================================================== 00:33:56.831 [2024-11-26T18:30:09.456Z] Total : 182601.70 713.29 0.00 0.00 697.35 298.67 1966.08 00:33:56.831 8572.00 IOPS, 33.48 MiB/s 00:33:56.831 Latency(us) 00:33:56.831 [2024-11-26T18:30:09.456Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:56.831 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:33:56.831 Nvme1n1 : 1.02 8594.02 33.57 0.00 0.00 14835.64 2088.96 23920.64 00:33:56.831 [2024-11-26T18:30:09.456Z] =================================================================================================================== 00:33:56.831 [2024-11-26T18:30:09.456Z] Total : 8594.02 33.57 0.00 0.00 14835.64 2088.96 23920.64 00:33:56.831 18995.00 IOPS, 74.20 MiB/s 00:33:56.831 Latency(us) 00:33:56.831 [2024-11-26T18:30:09.456Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:56.831 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:33:56.831 Nvme1n1 : 1.00 19037.71 74.37 0.00 0.00 6707.82 3017.39 10977.28 00:33:56.831 [2024-11-26T18:30:09.456Z] =================================================================================================================== 00:33:56.831 [2024-11-26T18:30:09.456Z] Total : 19037.71 74.37 0.00 0.00 6707.82 3017.39 10977.28 00:33:56.831 8628.00 IOPS, 33.70 MiB/s 00:33:56.831 Latency(us) 00:33:56.831 [2024-11-26T18:30:09.456Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:56.831 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:33:56.831 Nvme1n1 : 1.01 8750.70 34.18 0.00 0.00 14593.42 3467.95 30365.01 00:33:56.831 [2024-11-26T18:30:09.456Z] =================================================================================================================== 00:33:56.831 [2024-11-26T18:30:09.456Z] Total : 8750.70 34.18 0.00 0.00 14593.42 3467.95 30365.01 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 4120936 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 4120939 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 4120942 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # nvmfcleanup 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:33:56.831 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:33:56.831 rmmod nvme_tcp 00:33:56.831 rmmod nvme_fabrics 00:33:56.831 rmmod nvme_keyring 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@517 -- # '[' -n 4120735 ']' 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # killprocess 4120735 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # '[' -z 4120735 ']' 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@958 -- # kill -0 4120735 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # uname 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4120735 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4120735' 00:33:57.093 killing process with pid 4120735 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@973 -- # kill 4120735 00:33:57.093 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@978 -- # wait 4120735 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-save 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-restore 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:57.094 19:30:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:59.635 00:33:59.635 real 0m13.857s 00:33:59.635 user 0m15.365s 00:33:59.635 sys 0m8.106s 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:33:59.635 ************************************ 00:33:59.635 END TEST nvmf_bdev_io_wait 00:33:59.635 ************************************ 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:59.635 ************************************ 00:33:59.635 START TEST nvmf_queue_depth 00:33:59.635 ************************************ 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:33:59.635 * Looking for test storage... 00:33:59.635 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1693 -- # lcov --version 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:33:59.635 19:30:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:33:59.635 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:59.635 --rc genhtml_branch_coverage=1 00:33:59.635 --rc genhtml_function_coverage=1 00:33:59.635 --rc genhtml_legend=1 00:33:59.635 --rc geninfo_all_blocks=1 00:33:59.635 --rc geninfo_unexecuted_blocks=1 00:33:59.635 00:33:59.635 ' 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:33:59.635 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:59.635 --rc genhtml_branch_coverage=1 00:33:59.635 --rc genhtml_function_coverage=1 00:33:59.635 --rc genhtml_legend=1 00:33:59.635 --rc geninfo_all_blocks=1 00:33:59.635 --rc geninfo_unexecuted_blocks=1 00:33:59.635 00:33:59.635 ' 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:33:59.635 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:59.635 --rc genhtml_branch_coverage=1 00:33:59.635 --rc genhtml_function_coverage=1 00:33:59.635 --rc genhtml_legend=1 00:33:59.635 --rc geninfo_all_blocks=1 00:33:59.635 --rc geninfo_unexecuted_blocks=1 00:33:59.635 00:33:59.635 ' 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:33:59.635 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:59.635 --rc genhtml_branch_coverage=1 00:33:59.635 --rc genhtml_function_coverage=1 00:33:59.635 --rc genhtml_legend=1 00:33:59.635 --rc geninfo_all_blocks=1 00:33:59.635 --rc geninfo_unexecuted_blocks=1 00:33:59.635 00:33:59.635 ' 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:59.635 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@476 -- # prepare_net_devs 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@438 -- # local -g is_hw=no 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # remove_spdk_ns 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:33:59.636 19:30:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:34:07.774 Found 0000:31:00.0 (0x8086 - 0x159b) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:34:07.774 Found 0000:31:00.1 (0x8086 - 0x159b) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:07.774 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:34:07.775 Found net devices under 0000:31:00.0: cvl_0_0 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:34:07.775 Found net devices under 0000:31:00.1: cvl_0_1 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # is_hw=yes 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:07.775 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:08.037 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:08.037 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.575 ms 00:34:08.037 00:34:08.037 --- 10.0.0.2 ping statistics --- 00:34:08.037 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:08.037 rtt min/avg/max/mdev = 0.575/0.575/0.575/0.000 ms 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:08.037 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:08.037 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.275 ms 00:34:08.037 00:34:08.037 --- 10.0.0.1 ping statistics --- 00:34:08.037 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:08.037 rtt min/avg/max/mdev = 0.275/0.275/0.275/0.000 ms 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@450 -- # return 0 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:34:08.037 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@726 -- # xtrace_disable 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@509 -- # nvmfpid=4126452 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@510 -- # waitforlisten 4126452 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 4126452 ']' 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:08.038 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:08.038 19:30:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:08.038 [2024-11-26 19:30:20.554766] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:34:08.038 [2024-11-26 19:30:20.555925] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:34:08.038 [2024-11-26 19:30:20.555981] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:08.298 [2024-11-26 19:30:20.670885] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:08.298 [2024-11-26 19:30:20.721456] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:08.299 [2024-11-26 19:30:20.721512] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:08.299 [2024-11-26 19:30:20.721521] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:08.299 [2024-11-26 19:30:20.721529] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:08.299 [2024-11-26 19:30:20.721535] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:08.299 [2024-11-26 19:30:20.722334] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:08.299 [2024-11-26 19:30:20.798924] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:34:08.299 [2024-11-26 19:30:20.799217] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@732 -- # xtrace_disable 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:08.870 [2024-11-26 19:30:21.419232] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:08.870 Malloc0 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:08.870 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:09.131 [2024-11-26 19:30:21.499309] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:09.131 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:09.131 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=4126607 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 4126607 /var/tmp/bdevperf.sock 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 4126607 ']' 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:34:09.132 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:09.132 19:30:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:09.132 [2024-11-26 19:30:21.566392] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:34:09.132 [2024-11-26 19:30:21.566456] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4126607 ] 00:34:09.132 [2024-11-26 19:30:21.649691] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:09.132 [2024-11-26 19:30:21.691287] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:10.072 19:30:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:10.072 19:30:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:34:10.072 19:30:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:34:10.072 19:30:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.072 19:30:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:10.072 NVMe0n1 00:34:10.072 19:30:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.072 19:30:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:34:10.072 Running I/O for 10 seconds... 00:34:11.952 8758.00 IOPS, 34.21 MiB/s [2024-11-26T18:30:25.958Z] 9073.00 IOPS, 35.44 MiB/s [2024-11-26T18:30:26.529Z] 9216.00 IOPS, 36.00 MiB/s [2024-11-26T18:30:27.914Z] 9836.75 IOPS, 38.42 MiB/s [2024-11-26T18:30:28.857Z] 10244.00 IOPS, 40.02 MiB/s [2024-11-26T18:30:29.798Z] 10555.33 IOPS, 41.23 MiB/s [2024-11-26T18:30:30.758Z] 10710.57 IOPS, 41.84 MiB/s [2024-11-26T18:30:31.708Z] 10881.75 IOPS, 42.51 MiB/s [2024-11-26T18:30:32.652Z] 10968.78 IOPS, 42.85 MiB/s [2024-11-26T18:30:32.652Z] 11062.50 IOPS, 43.21 MiB/s 00:34:20.027 Latency(us) 00:34:20.027 [2024-11-26T18:30:32.652Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:20.027 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:34:20.027 Verification LBA range: start 0x0 length 0x4000 00:34:20.027 NVMe0n1 : 10.06 11098.50 43.35 0.00 0.00 91971.33 21626.88 69905.07 00:34:20.027 [2024-11-26T18:30:32.652Z] =================================================================================================================== 00:34:20.027 [2024-11-26T18:30:32.652Z] Total : 11098.50 43.35 0.00 0.00 91971.33 21626.88 69905.07 00:34:20.027 { 00:34:20.027 "results": [ 00:34:20.027 { 00:34:20.027 "job": "NVMe0n1", 00:34:20.027 "core_mask": "0x1", 00:34:20.027 "workload": "verify", 00:34:20.027 "status": "finished", 00:34:20.027 "verify_range": { 00:34:20.027 "start": 0, 00:34:20.027 "length": 16384 00:34:20.027 }, 00:34:20.027 "queue_depth": 1024, 00:34:20.027 "io_size": 4096, 00:34:20.027 "runtime": 10.059825, 00:34:20.027 "iops": 11098.503204578608, 00:34:20.027 "mibps": 43.35352814288519, 00:34:20.027 "io_failed": 0, 00:34:20.027 "io_timeout": 0, 00:34:20.027 "avg_latency_us": 91971.32550761761, 00:34:20.027 "min_latency_us": 21626.88, 00:34:20.027 "max_latency_us": 69905.06666666667 00:34:20.027 } 00:34:20.027 ], 00:34:20.027 "core_count": 1 00:34:20.027 } 00:34:20.027 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 4126607 00:34:20.027 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 4126607 ']' 00:34:20.027 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 4126607 00:34:20.027 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:34:20.027 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:20.027 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4126607 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4126607' 00:34:20.287 killing process with pid 4126607 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 4126607 00:34:20.287 Received shutdown signal, test time was about 10.000000 seconds 00:34:20.287 00:34:20.287 Latency(us) 00:34:20.287 [2024-11-26T18:30:32.912Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:20.287 [2024-11-26T18:30:32.912Z] =================================================================================================================== 00:34:20.287 [2024-11-26T18:30:32.912Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 4126607 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@516 -- # nvmfcleanup 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:20.287 rmmod nvme_tcp 00:34:20.287 rmmod nvme_fabrics 00:34:20.287 rmmod nvme_keyring 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@517 -- # '[' -n 4126452 ']' 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@518 -- # killprocess 4126452 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 4126452 ']' 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 4126452 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:20.287 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4126452 00:34:20.548 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:34:20.548 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:34:20.548 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4126452' 00:34:20.548 killing process with pid 4126452 00:34:20.548 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 4126452 00:34:20.548 19:30:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 4126452 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-save 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-restore 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:20.548 19:30:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:23.096 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:23.096 00:34:23.096 real 0m23.341s 00:34:23.096 user 0m24.836s 00:34:23.096 sys 0m8.054s 00:34:23.096 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:23.096 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:34:23.096 ************************************ 00:34:23.096 END TEST nvmf_queue_depth 00:34:23.096 ************************************ 00:34:23.096 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:34:23.096 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:34:23.097 ************************************ 00:34:23.097 START TEST nvmf_target_multipath 00:34:23.097 ************************************ 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:34:23.097 * Looking for test storage... 00:34:23.097 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1693 -- # lcov --version 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:34:23.097 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:34:23.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:23.098 --rc genhtml_branch_coverage=1 00:34:23.098 --rc genhtml_function_coverage=1 00:34:23.098 --rc genhtml_legend=1 00:34:23.098 --rc geninfo_all_blocks=1 00:34:23.098 --rc geninfo_unexecuted_blocks=1 00:34:23.098 00:34:23.098 ' 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:34:23.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:23.098 --rc genhtml_branch_coverage=1 00:34:23.098 --rc genhtml_function_coverage=1 00:34:23.098 --rc genhtml_legend=1 00:34:23.098 --rc geninfo_all_blocks=1 00:34:23.098 --rc geninfo_unexecuted_blocks=1 00:34:23.098 00:34:23.098 ' 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:34:23.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:23.098 --rc genhtml_branch_coverage=1 00:34:23.098 --rc genhtml_function_coverage=1 00:34:23.098 --rc genhtml_legend=1 00:34:23.098 --rc geninfo_all_blocks=1 00:34:23.098 --rc geninfo_unexecuted_blocks=1 00:34:23.098 00:34:23.098 ' 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:34:23.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:23.098 --rc genhtml_branch_coverage=1 00:34:23.098 --rc genhtml_function_coverage=1 00:34:23.098 --rc genhtml_legend=1 00:34:23.098 --rc geninfo_all_blocks=1 00:34:23.098 --rc geninfo_unexecuted_blocks=1 00:34:23.098 00:34:23.098 ' 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:23.098 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:23.099 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:23.100 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:23.100 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:23.100 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:23.100 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:34:23.100 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:23.101 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:34:23.101 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:23.101 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@476 -- # prepare_net_devs 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@438 -- # local -g is_hw=no 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # remove_spdk_ns 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:23.102 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:23.103 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:34:23.103 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:34:23.103 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:34:23.103 19:30:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:34:31.251 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:34:31.252 Found 0000:31:00.0 (0x8086 - 0x159b) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:34:31.252 Found 0000:31:00.1 (0x8086 - 0x159b) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:34:31.252 Found net devices under 0000:31:00.0: cvl_0_0 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:34:31.252 Found net devices under 0000:31:00.1: cvl_0_1 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # is_hw=yes 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:31.252 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:31.253 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:31.253 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.758 ms 00:34:31.253 00:34:31.253 --- 10.0.0.2 ping statistics --- 00:34:31.253 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:31.253 rtt min/avg/max/mdev = 0.758/0.758/0.758/0.000 ms 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:31.253 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:31.253 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.217 ms 00:34:31.253 00:34:31.253 --- 10.0.0.1 ping statistics --- 00:34:31.253 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:31.253 rtt min/avg/max/mdev = 0.217/0.217/0.217/0.000 ms 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@450 -- # return 0 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:34:31.253 only one NIC for nvmf test 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:31.253 rmmod nvme_tcp 00:34:31.253 rmmod nvme_fabrics 00:34:31.253 rmmod nvme_keyring 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:31.253 19:30:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:34:33.167 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:33.168 00:34:33.168 real 0m10.536s 00:34:33.168 user 0m2.322s 00:34:33.168 sys 0m6.118s 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:33.168 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:34:33.168 ************************************ 00:34:33.168 END TEST nvmf_target_multipath 00:34:33.168 ************************************ 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:34:33.430 ************************************ 00:34:33.430 START TEST nvmf_zcopy 00:34:33.430 ************************************ 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:34:33.430 * Looking for test storage... 00:34:33.430 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1693 -- # lcov --version 00:34:33.430 19:30:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:34:33.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:33.430 --rc genhtml_branch_coverage=1 00:34:33.430 --rc genhtml_function_coverage=1 00:34:33.430 --rc genhtml_legend=1 00:34:33.430 --rc geninfo_all_blocks=1 00:34:33.430 --rc geninfo_unexecuted_blocks=1 00:34:33.430 00:34:33.430 ' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:34:33.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:33.430 --rc genhtml_branch_coverage=1 00:34:33.430 --rc genhtml_function_coverage=1 00:34:33.430 --rc genhtml_legend=1 00:34:33.430 --rc geninfo_all_blocks=1 00:34:33.430 --rc geninfo_unexecuted_blocks=1 00:34:33.430 00:34:33.430 ' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:34:33.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:33.430 --rc genhtml_branch_coverage=1 00:34:33.430 --rc genhtml_function_coverage=1 00:34:33.430 --rc genhtml_legend=1 00:34:33.430 --rc geninfo_all_blocks=1 00:34:33.430 --rc geninfo_unexecuted_blocks=1 00:34:33.430 00:34:33.430 ' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:34:33.430 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:33.430 --rc genhtml_branch_coverage=1 00:34:33.430 --rc genhtml_function_coverage=1 00:34:33.430 --rc genhtml_legend=1 00:34:33.430 --rc geninfo_all_blocks=1 00:34:33.430 --rc geninfo_unexecuted_blocks=1 00:34:33.430 00:34:33.430 ' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:33.430 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@476 -- # prepare_net_devs 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@438 -- # local -g is_hw=no 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # remove_spdk_ns 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:34:33.431 19:30:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:34:41.572 Found 0000:31:00.0 (0x8086 - 0x159b) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:34:41.572 Found 0000:31:00.1 (0x8086 - 0x159b) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:34:41.572 Found net devices under 0000:31:00.0: cvl_0_0 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:34:41.572 Found net devices under 0000:31:00.1: cvl_0_1 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # is_hw=yes 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:41.572 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:41.573 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:41.573 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.688 ms 00:34:41.573 00:34:41.573 --- 10.0.0.2 ping statistics --- 00:34:41.573 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:41.573 rtt min/avg/max/mdev = 0.688/0.688/0.688/0.000 ms 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:41.573 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:41.573 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.326 ms 00:34:41.573 00:34:41.573 --- 10.0.0.1 ping statistics --- 00:34:41.573 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:41.573 rtt min/avg/max/mdev = 0.326/0.326/0.326/0.000 ms 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@450 -- # return 0 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:34:41.573 19:30:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@726 -- # xtrace_disable 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@509 -- # nvmfpid=4137966 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@510 -- # waitforlisten 4137966 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@835 -- # '[' -z 4137966 ']' 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:41.573 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:41.573 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:41.573 [2024-11-26 19:30:54.061208] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:34:41.573 [2024-11-26 19:30:54.062210] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:34:41.573 [2024-11-26 19:30:54.062250] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:41.573 [2024-11-26 19:30:54.165008] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:41.832 [2024-11-26 19:30:54.207576] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:41.832 [2024-11-26 19:30:54.207624] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:41.832 [2024-11-26 19:30:54.207633] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:41.832 [2024-11-26 19:30:54.207640] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:41.832 [2024-11-26 19:30:54.207646] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:41.832 [2024-11-26 19:30:54.208339] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:41.832 [2024-11-26 19:30:54.279657] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:34:41.832 [2024-11-26 19:30:54.279953] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@868 -- # return 0 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@732 -- # xtrace_disable 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:42.402 [2024-11-26 19:30:54.917203] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:42.402 [2024-11-26 19:30:54.945510] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:42.402 malloc0 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:34:42.402 { 00:34:42.402 "params": { 00:34:42.402 "name": "Nvme$subsystem", 00:34:42.402 "trtype": "$TEST_TRANSPORT", 00:34:42.402 "traddr": "$NVMF_FIRST_TARGET_IP", 00:34:42.402 "adrfam": "ipv4", 00:34:42.402 "trsvcid": "$NVMF_PORT", 00:34:42.402 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:34:42.402 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:34:42.402 "hdgst": ${hdgst:-false}, 00:34:42.402 "ddgst": ${ddgst:-false} 00:34:42.402 }, 00:34:42.402 "method": "bdev_nvme_attach_controller" 00:34:42.402 } 00:34:42.402 EOF 00:34:42.402 )") 00:34:42.402 19:30:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:34:42.402 19:30:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:34:42.402 19:30:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:34:42.402 19:30:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:34:42.402 "params": { 00:34:42.402 "name": "Nvme1", 00:34:42.402 "trtype": "tcp", 00:34:42.402 "traddr": "10.0.0.2", 00:34:42.402 "adrfam": "ipv4", 00:34:42.402 "trsvcid": "4420", 00:34:42.402 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:34:42.402 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:34:42.402 "hdgst": false, 00:34:42.402 "ddgst": false 00:34:42.402 }, 00:34:42.402 "method": "bdev_nvme_attach_controller" 00:34:42.402 }' 00:34:42.662 [2024-11-26 19:30:55.045170] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:34:42.662 [2024-11-26 19:30:55.045225] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4138154 ] 00:34:42.662 [2024-11-26 19:30:55.123940] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:42.662 [2024-11-26 19:30:55.161455] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:42.922 Running I/O for 10 seconds... 00:34:44.802 6557.00 IOPS, 51.23 MiB/s [2024-11-26T18:30:58.369Z] 6595.50 IOPS, 51.53 MiB/s [2024-11-26T18:30:59.750Z] 6612.33 IOPS, 51.66 MiB/s [2024-11-26T18:31:00.691Z] 6620.50 IOPS, 51.72 MiB/s [2024-11-26T18:31:01.630Z] 6620.20 IOPS, 51.72 MiB/s [2024-11-26T18:31:02.573Z] 6961.33 IOPS, 54.39 MiB/s [2024-11-26T18:31:03.511Z] 7334.29 IOPS, 57.30 MiB/s [2024-11-26T18:31:04.450Z] 7615.50 IOPS, 59.50 MiB/s [2024-11-26T18:31:05.392Z] 7834.44 IOPS, 61.21 MiB/s [2024-11-26T18:31:05.392Z] 8012.20 IOPS, 62.60 MiB/s 00:34:52.767 Latency(us) 00:34:52.767 [2024-11-26T18:31:05.392Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:52.767 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:34:52.767 Verification LBA range: start 0x0 length 0x1000 00:34:52.767 Nvme1n1 : 10.01 8014.52 62.61 0.00 0.00 15918.00 1761.28 27088.21 00:34:52.767 [2024-11-26T18:31:05.392Z] =================================================================================================================== 00:34:52.767 [2024-11-26T18:31:05.392Z] Total : 8014.52 62.61 0.00 0.00 15918.00 1761.28 27088.21 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=4140055 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:34:53.028 { 00:34:53.028 "params": { 00:34:53.028 "name": "Nvme$subsystem", 00:34:53.028 "trtype": "$TEST_TRANSPORT", 00:34:53.028 "traddr": "$NVMF_FIRST_TARGET_IP", 00:34:53.028 "adrfam": "ipv4", 00:34:53.028 "trsvcid": "$NVMF_PORT", 00:34:53.028 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:34:53.028 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:34:53.028 "hdgst": ${hdgst:-false}, 00:34:53.028 "ddgst": ${ddgst:-false} 00:34:53.028 }, 00:34:53.028 "method": "bdev_nvme_attach_controller" 00:34:53.028 } 00:34:53.028 EOF 00:34:53.028 )") 00:34:53.028 [2024-11-26 19:31:05.500731] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.500761] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:34:53.028 19:31:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:34:53.028 "params": { 00:34:53.028 "name": "Nvme1", 00:34:53.028 "trtype": "tcp", 00:34:53.028 "traddr": "10.0.0.2", 00:34:53.028 "adrfam": "ipv4", 00:34:53.028 "trsvcid": "4420", 00:34:53.028 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:34:53.028 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:34:53.028 "hdgst": false, 00:34:53.028 "ddgst": false 00:34:53.028 }, 00:34:53.028 "method": "bdev_nvme_attach_controller" 00:34:53.028 }' 00:34:53.028 [2024-11-26 19:31:05.512690] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.512700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.524687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.524695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.536688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.536696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.544257] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:34:53.028 [2024-11-26 19:31:05.544304] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4140055 ] 00:34:53.028 [2024-11-26 19:31:05.548688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.548696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.560687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.560695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.572687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.572695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.584687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.584695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.596687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.596694] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.608687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.608694] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.620372] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:53.028 [2024-11-26 19:31:05.620687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.620694] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.632689] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.632698] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.028 [2024-11-26 19:31:05.644695] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.028 [2024-11-26 19:31:05.644705] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.655830] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:53.294 [2024-11-26 19:31:05.656688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.656696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.668692] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.668701] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.680693] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.680705] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.692691] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.692703] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.704691] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.704700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.716688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.716696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.728699] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.728714] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.740695] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.740708] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.752692] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.752702] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.764696] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.764709] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.776689] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.776696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.788688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.788697] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.800690] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.800701] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.812690] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.812699] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.824689] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.294 [2024-11-26 19:31:05.824697] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.294 [2024-11-26 19:31:05.836688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.295 [2024-11-26 19:31:05.836695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.295 [2024-11-26 19:31:05.848687] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.295 [2024-11-26 19:31:05.848695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.295 [2024-11-26 19:31:05.860689] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.295 [2024-11-26 19:31:05.860698] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.295 [2024-11-26 19:31:05.872688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.295 [2024-11-26 19:31:05.872695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.295 [2024-11-26 19:31:05.884688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.295 [2024-11-26 19:31:05.884696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.295 [2024-11-26 19:31:05.896690] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.295 [2024-11-26 19:31:05.896700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.295 [2024-11-26 19:31:05.908689] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.295 [2024-11-26 19:31:05.908696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:05.920688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:05.920696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:05.932689] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:05.932695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:05.944691] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:05.944702] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:05.986296] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:05.986308] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:05.996690] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:05.996700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 Running I/O for 5 seconds... 00:34:53.634 [2024-11-26 19:31:06.012075] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.012092] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.025252] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.025268] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.039769] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.039785] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.052650] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.052666] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.065688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.065703] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.079886] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.079901] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.092995] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.093010] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.107759] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.107774] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.120800] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.120814] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.133843] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.133861] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.148085] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.148100] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.160981] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.160996] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.175674] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.175689] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.188940] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.188956] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.201843] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.201865] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.215950] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.215964] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.228914] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.228928] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.634 [2024-11-26 19:31:06.241622] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.634 [2024-11-26 19:31:06.241636] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.256042] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.256057] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.269318] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.269332] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.284417] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.284432] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.297639] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.297653] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.312099] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.312114] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.325300] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.325314] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.340216] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.340231] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.353394] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.353409] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.367522] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.367536] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.380524] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.380539] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.393296] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.393310] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.408010] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.408024] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.421081] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.421095] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.435897] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.435911] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.449119] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.449133] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.464139] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.464158] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.477323] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.477337] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.941 [2024-11-26 19:31:06.491675] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.941 [2024-11-26 19:31:06.491689] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.942 [2024-11-26 19:31:06.504410] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.942 [2024-11-26 19:31:06.504424] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.942 [2024-11-26 19:31:06.517775] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.942 [2024-11-26 19:31:06.517789] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.942 [2024-11-26 19:31:06.532364] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.942 [2024-11-26 19:31:06.532379] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.942 [2024-11-26 19:31:06.545336] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.942 [2024-11-26 19:31:06.545351] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:53.942 [2024-11-26 19:31:06.559721] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:53.942 [2024-11-26 19:31:06.559735] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.572638] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.572652] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.585368] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.585383] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.599877] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.599892] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.612638] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.612653] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.625378] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.625393] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.639461] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.639476] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.652345] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.652359] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.665297] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.665311] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.680207] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.680222] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.693427] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.693441] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.707659] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.707674] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.720439] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.720461] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.733568] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.733582] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.747682] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.747696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.760798] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.760812] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.773525] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.773539] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.788157] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.788171] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.801180] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.801194] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.203 [2024-11-26 19:31:06.815847] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.203 [2024-11-26 19:31:06.815865] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.828874] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.828889] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.841849] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.841867] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.855668] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.855683] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.868693] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.868707] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.881499] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.881513] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.895848] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.895867] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.908847] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.908866] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.921841] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.921856] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.936038] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.936052] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.948916] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.948931] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.961365] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.961379] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.975903] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.975918] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:06.989068] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:06.989082] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:07.003519] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:07.003534] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 19082.00 IOPS, 149.08 MiB/s [2024-11-26T18:31:07.089Z] [2024-11-26 19:31:07.016312] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:07.016326] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:07.029754] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:07.029768] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:07.043477] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:07.043491] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:07.056554] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:07.056568] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:07.069854] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:07.069874] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.464 [2024-11-26 19:31:07.083830] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.464 [2024-11-26 19:31:07.083846] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.096710] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.096725] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.109733] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.109747] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.124464] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.124479] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.137690] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.137704] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.152752] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.152767] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.166036] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.166050] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.179952] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.179967] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.192742] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.192757] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.205472] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.205486] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.220243] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.220258] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.233493] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.233507] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.247803] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.247818] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.260858] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.260879] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.273622] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.273636] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.287794] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.287809] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.300727] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.300741] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.313906] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.313920] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.328456] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.328471] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.726 [2024-11-26 19:31:07.341554] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.726 [2024-11-26 19:31:07.341568] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.355763] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.355778] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.368741] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.368756] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.381714] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.381731] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.396077] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.396092] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.408882] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.408897] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.421867] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.421882] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.436248] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.436264] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.449198] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.449213] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.463774] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.463789] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.476818] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.476836] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.489583] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.489597] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.504093] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.504107] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.517125] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.517139] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.532134] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.532149] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.545263] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.545277] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.560167] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.560182] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.573243] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.573257] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.587915] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.587930] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:54.987 [2024-11-26 19:31:07.601059] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:54.987 [2024-11-26 19:31:07.601073] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.248 [2024-11-26 19:31:07.615874] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.248 [2024-11-26 19:31:07.615890] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.248 [2024-11-26 19:31:07.629149] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.248 [2024-11-26 19:31:07.629163] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.248 [2024-11-26 19:31:07.644011] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.248 [2024-11-26 19:31:07.644026] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.248 [2024-11-26 19:31:07.657079] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.248 [2024-11-26 19:31:07.657093] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.248 [2024-11-26 19:31:07.671921] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.248 [2024-11-26 19:31:07.671936] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.248 [2024-11-26 19:31:07.685221] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.248 [2024-11-26 19:31:07.685236] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.248 [2024-11-26 19:31:07.699682] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.699696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.712691] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.712705] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.725399] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.725413] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.739776] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.739796] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.753040] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.753054] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.767350] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.767365] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.780869] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.780884] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.793843] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.793858] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.808258] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.808273] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.821174] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.821187] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.835823] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.835838] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.849159] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.849174] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.249 [2024-11-26 19:31:07.863990] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.249 [2024-11-26 19:31:07.864005] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.877194] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.877208] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.891771] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.891786] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.905079] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.905092] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.919572] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.919586] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.932498] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.932513] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.945462] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.945476] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.960556] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.960571] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.973565] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.973579] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:07.987744] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:07.987758] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.000896] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.000914] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 19071.00 IOPS, 148.99 MiB/s [2024-11-26T18:31:08.135Z] [2024-11-26 19:31:08.011912] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.011927] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.024747] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.024761] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.037811] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.037825] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.051921] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.051936] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.065294] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.065308] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.079666] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.079681] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.092567] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.092581] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.105466] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.105480] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.510 [2024-11-26 19:31:08.120139] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.510 [2024-11-26 19:31:08.120154] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.133106] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.133121] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.148249] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.148263] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.161541] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.161556] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.175551] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.175565] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.188666] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.188680] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.201548] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.201562] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.216704] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.216719] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.229555] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.229569] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.244036] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.244051] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.256907] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.256922] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.270154] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.270168] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.284375] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.284390] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.297383] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.297397] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.311753] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.311768] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.324742] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.324757] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.337163] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.337177] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.351664] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.351679] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.364942] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.364957] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.377812] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.377827] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:55.772 [2024-11-26 19:31:08.391693] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:55.772 [2024-11-26 19:31:08.391708] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.404666] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.404681] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.417894] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.417909] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.431845] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.431859] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.444943] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.444958] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.457706] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.457720] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.472090] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.472104] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.485135] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.485149] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.499785] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.499800] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.512680] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.512694] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.525441] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.525455] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.540026] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.540040] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.552826] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.552841] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.565732] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.565747] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.580147] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.580161] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.593266] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.593280] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.607161] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.607176] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.620380] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.620394] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.633034] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.633049] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.034 [2024-11-26 19:31:08.648276] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.034 [2024-11-26 19:31:08.648290] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.661140] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.661154] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.676102] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.676116] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.689235] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.689249] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.703799] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.703813] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.716679] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.716693] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.729852] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.729870] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.743753] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.743767] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.756577] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.756592] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.769932] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.769946] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.783908] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.783923] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.796758] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.796773] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.809405] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.809419] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.824269] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.824284] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.837254] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.837268] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.852013] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.852027] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.865005] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.865020] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.879989] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.880004] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.893007] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.893020] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.295 [2024-11-26 19:31:08.907845] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.295 [2024-11-26 19:31:08.907860] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:08.920848] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:08.920868] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:08.933792] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:08.933807] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:08.948124] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:08.948139] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:08.961262] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:08.961276] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:08.975567] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:08.975581] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:08.988611] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:08.988626] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.002117] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.002132] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 19093.33 IOPS, 149.17 MiB/s [2024-11-26T18:31:09.181Z] [2024-11-26 19:31:09.015943] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.015963] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.028902] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.028917] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.041845] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.041859] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.056396] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.056411] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.069470] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.069484] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.084315] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.084329] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.097279] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.097293] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.111780] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.111795] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.125019] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.556 [2024-11-26 19:31:09.125034] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.556 [2024-11-26 19:31:09.139722] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.557 [2024-11-26 19:31:09.139738] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.557 [2024-11-26 19:31:09.152729] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.557 [2024-11-26 19:31:09.152743] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.557 [2024-11-26 19:31:09.166067] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.557 [2024-11-26 19:31:09.166081] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.179905] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.179920] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.192898] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.192913] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.205895] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.205909] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.219911] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.219926] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.233105] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.233119] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.248003] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.248018] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.261058] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.261072] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.275542] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.275561] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.288433] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.288447] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.301328] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.301342] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.315824] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.315839] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.328731] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.328746] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.341355] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.341369] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.355659] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.355674] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.368853] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.368872] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.382002] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.382017] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.396314] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.396328] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.409271] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.409286] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.423798] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.423813] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:56.817 [2024-11-26 19:31:09.436992] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:56.817 [2024-11-26 19:31:09.437006] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.452455] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.452471] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.465846] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.465866] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.479664] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.479678] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.492524] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.492538] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.505169] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.505184] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.519760] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.519775] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.533008] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.533026] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.547756] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.547770] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.560734] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.560749] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.573873] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.573888] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.588528] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.588542] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.601394] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.601408] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.615802] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.615816] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.628826] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.628841] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.641600] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.641615] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.656351] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.656366] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.669051] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.669065] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.683945] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.683959] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.078 [2024-11-26 19:31:09.696654] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.078 [2024-11-26 19:31:09.696669] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.709783] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.709797] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.723801] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.723815] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.736791] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.736805] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.749790] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.749804] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.763684] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.763698] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.776860] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.776878] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.789905] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.789923] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.804484] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.804498] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.817334] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.817348] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.831773] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.831787] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.844781] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.844796] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.857994] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.858008] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.872488] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.872502] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.885628] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.885642] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.899659] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.899673] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.912490] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.912505] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.925622] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.925636] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.939950] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.939965] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.339 [2024-11-26 19:31:09.952607] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.339 [2024-11-26 19:31:09.952622] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:09.965597] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:09.965611] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:09.979567] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:09.979582] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:09.992493] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:09.992507] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.006463] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.006479] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 19078.50 IOPS, 149.05 MiB/s [2024-11-26T18:31:10.225Z] [2024-11-26 19:31:10.020532] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.020548] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.033660] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.033674] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.048115] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.048133] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.061176] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.061190] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.076022] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.076037] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.089152] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.089166] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.103965] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.103980] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.117260] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.117274] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.131728] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.131743] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.145140] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.145154] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.159645] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.159659] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.172692] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.172706] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.185777] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.185791] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.199985] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.199999] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.600 [2024-11-26 19:31:10.213075] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.600 [2024-11-26 19:31:10.213089] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.227436] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.227451] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.240589] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.240604] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.253715] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.253729] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.267526] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.267540] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.280941] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.280956] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.293834] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.293847] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.307691] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.307706] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.320879] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.320894] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.333560] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.333574] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.347762] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.347777] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.360522] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.360536] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.374103] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.374117] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.388114] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.388129] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.401242] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.401256] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.416280] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.416295] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.429285] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.429299] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.443760] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.443775] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.456553] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.456568] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.469057] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.469071] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:57.861 [2024-11-26 19:31:10.483539] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:57.861 [2024-11-26 19:31:10.483553] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.496633] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.496648] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.510034] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.510048] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.524162] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.524177] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.536998] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.537012] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.551821] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.551836] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.564872] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.564887] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.577994] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.578008] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.591961] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.591976] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.604727] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.604742] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.617446] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.617459] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.632046] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.632059] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.644916] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.121 [2024-11-26 19:31:10.644930] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.121 [2024-11-26 19:31:10.658183] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.122 [2024-11-26 19:31:10.658197] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.122 [2024-11-26 19:31:10.671953] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.122 [2024-11-26 19:31:10.671968] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.122 [2024-11-26 19:31:10.685105] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.122 [2024-11-26 19:31:10.685119] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.122 [2024-11-26 19:31:10.699808] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.122 [2024-11-26 19:31:10.699824] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.122 [2024-11-26 19:31:10.713161] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.122 [2024-11-26 19:31:10.713175] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.122 [2024-11-26 19:31:10.727851] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.122 [2024-11-26 19:31:10.727870] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.122 [2024-11-26 19:31:10.740745] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.122 [2024-11-26 19:31:10.740759] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.753564] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.753579] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.767985] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.768000] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.781153] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.781168] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.795439] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.795454] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.808601] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.808621] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.821434] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.821448] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.835558] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.835573] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.848246] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.848262] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.861136] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.861150] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.875910] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.875925] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.888830] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.888844] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.901402] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.901417] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.915840] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.915854] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.928902] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.928917] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.941761] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.941776] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.955725] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.955740] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.968949] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.968964] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.981820] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.981834] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.382 [2024-11-26 19:31:10.995848] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.382 [2024-11-26 19:31:10.995868] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.008724] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.008739] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 19089.40 IOPS, 149.14 MiB/s [2024-11-26T18:31:11.267Z] [2024-11-26 19:31:11.016695] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.016710] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 00:34:58.642 Latency(us) 00:34:58.642 [2024-11-26T18:31:11.267Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:58.642 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:34:58.642 Nvme1n1 : 5.00 19089.52 149.14 0.00 0.00 6699.18 2594.13 13489.49 00:34:58.642 [2024-11-26T18:31:11.267Z] =================================================================================================================== 00:34:58.642 [2024-11-26T18:31:11.267Z] Total : 19089.52 149.14 0.00 0.00 6699.18 2594.13 13489.49 00:34:58.642 [2024-11-26 19:31:11.028692] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.028705] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.040698] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.040709] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.052693] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.052705] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.064692] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.064703] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.076690] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.076700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.088688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.088698] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.100688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.100695] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.112691] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.112702] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.124695] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.124708] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 [2024-11-26 19:31:11.136688] subsystem.c:2126:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:34:58.642 [2024-11-26 19:31:11.136696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:34:58.642 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (4140055) - No such process 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 4140055 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:58.642 delay0 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:34:58.642 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:58.643 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:34:58.643 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:58.643 19:31:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:34:58.902 [2024-11-26 19:31:11.283231] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:35:05.480 [2024-11-26 19:31:17.751123] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x82eb10 is same with the state(6) to be set 00:35:05.480 Initializing NVMe Controllers 00:35:05.480 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:35:05.480 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:35:05.480 Initialization complete. Launching workers. 00:35:05.480 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 2141 00:35:05.480 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 2425, failed to submit 36 00:35:05.480 success 2242, unsuccessful 183, failed 0 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@516 -- # nvmfcleanup 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:05.480 rmmod nvme_tcp 00:35:05.480 rmmod nvme_fabrics 00:35:05.480 rmmod nvme_keyring 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@517 -- # '[' -n 4137966 ']' 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@518 -- # killprocess 4137966 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@954 -- # '[' -z 4137966 ']' 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@958 -- # kill -0 4137966 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@959 -- # uname 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4137966 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4137966' 00:35:05.480 killing process with pid 4137966 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@973 -- # kill 4137966 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@978 -- # wait 4137966 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:35:05.480 19:31:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-save 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-restore 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:05.480 19:31:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:08.030 00:35:08.030 real 0m34.240s 00:35:08.030 user 0m43.565s 00:35:08.030 sys 0m12.171s 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:35:08.030 ************************************ 00:35:08.030 END TEST nvmf_zcopy 00:35:08.030 ************************************ 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:35:08.030 ************************************ 00:35:08.030 START TEST nvmf_nmic 00:35:08.030 ************************************ 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:35:08.030 * Looking for test storage... 00:35:08.030 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1693 -- # lcov --version 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:35:08.030 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.030 --rc genhtml_branch_coverage=1 00:35:08.030 --rc genhtml_function_coverage=1 00:35:08.030 --rc genhtml_legend=1 00:35:08.030 --rc geninfo_all_blocks=1 00:35:08.030 --rc geninfo_unexecuted_blocks=1 00:35:08.030 00:35:08.030 ' 00:35:08.030 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:35:08.031 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.031 --rc genhtml_branch_coverage=1 00:35:08.031 --rc genhtml_function_coverage=1 00:35:08.031 --rc genhtml_legend=1 00:35:08.031 --rc geninfo_all_blocks=1 00:35:08.031 --rc geninfo_unexecuted_blocks=1 00:35:08.031 00:35:08.031 ' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:35:08.031 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.031 --rc genhtml_branch_coverage=1 00:35:08.031 --rc genhtml_function_coverage=1 00:35:08.031 --rc genhtml_legend=1 00:35:08.031 --rc geninfo_all_blocks=1 00:35:08.031 --rc geninfo_unexecuted_blocks=1 00:35:08.031 00:35:08.031 ' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:35:08.031 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.031 --rc genhtml_branch_coverage=1 00:35:08.031 --rc genhtml_function_coverage=1 00:35:08.031 --rc genhtml_legend=1 00:35:08.031 --rc geninfo_all_blocks=1 00:35:08.031 --rc geninfo_unexecuted_blocks=1 00:35:08.031 00:35:08.031 ' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@476 -- # prepare_net_devs 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@438 -- # local -g is_hw=no 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # remove_spdk_ns 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:35:08.031 19:31:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:35:16.195 Found 0000:31:00.0 (0x8086 - 0x159b) 00:35:16.195 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:35:16.196 Found 0000:31:00.1 (0x8086 - 0x159b) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:35:16.196 Found net devices under 0000:31:00.0: cvl_0_0 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:35:16.196 Found net devices under 0000:31:00.1: cvl_0_1 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # is_hw=yes 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:16.196 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:16.458 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:16.458 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.685 ms 00:35:16.458 00:35:16.458 --- 10.0.0.2 ping statistics --- 00:35:16.458 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:16.458 rtt min/avg/max/mdev = 0.685/0.685/0.685/0.000 ms 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:16.458 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:16.458 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.281 ms 00:35:16.458 00:35:16.458 --- 10.0.0.1 ping statistics --- 00:35:16.458 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:16.458 rtt min/avg/max/mdev = 0.281/0.281/0.281/0.000 ms 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@450 -- # return 0 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@509 -- # nvmfpid=4147048 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@510 -- # waitforlisten 4147048 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@835 -- # '[' -z 4147048 ']' 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:16.458 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:16.458 19:31:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:16.458 [2024-11-26 19:31:28.967287] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:35:16.458 [2024-11-26 19:31:28.968460] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:35:16.458 [2024-11-26 19:31:28.968510] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:16.458 [2024-11-26 19:31:29.064112] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:35:16.719 [2024-11-26 19:31:29.106922] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:16.719 [2024-11-26 19:31:29.106958] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:16.719 [2024-11-26 19:31:29.106967] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:16.719 [2024-11-26 19:31:29.106973] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:16.719 [2024-11-26 19:31:29.106979] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:16.719 [2024-11-26 19:31:29.108480] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:16.719 [2024-11-26 19:31:29.108599] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:35:16.719 [2024-11-26 19:31:29.108758] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:16.719 [2024-11-26 19:31:29.108759] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:35:16.719 [2024-11-26 19:31:29.165889] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:35:16.719 [2024-11-26 19:31:29.166094] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:35:16.719 [2024-11-26 19:31:29.167107] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:35:16.719 [2024-11-26 19:31:29.167743] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:35:16.719 [2024-11-26 19:31:29.167847] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@868 -- # return 0 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.290 [2024-11-26 19:31:29.833527] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.290 Malloc0 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.290 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.290 [2024-11-26 19:31:29.913415] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:17.551 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:35:17.552 test case1: single bdev can't be used in multiple subsystems 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.552 [2024-11-26 19:31:29.949148] bdev.c:8507:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:35:17.552 [2024-11-26 19:31:29.949167] subsystem.c:2156:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:35:17.552 [2024-11-26 19:31:29.949175] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:35:17.552 request: 00:35:17.552 { 00:35:17.552 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:35:17.552 "namespace": { 00:35:17.552 "bdev_name": "Malloc0", 00:35:17.552 "no_auto_visible": false, 00:35:17.552 "hide_metadata": false 00:35:17.552 }, 00:35:17.552 "method": "nvmf_subsystem_add_ns", 00:35:17.552 "req_id": 1 00:35:17.552 } 00:35:17.552 Got JSON-RPC error response 00:35:17.552 response: 00:35:17.552 { 00:35:17.552 "code": -32602, 00:35:17.552 "message": "Invalid parameters" 00:35:17.552 } 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:35:17.552 Adding namespace failed - expected result. 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:35:17.552 test case2: host connect to nvmf target in multiple paths 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:17.552 [2024-11-26 19:31:29.961254] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.552 19:31:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:35:17.811 19:31:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:35:18.381 19:31:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:35:18.381 19:31:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1202 -- # local i=0 00:35:18.381 19:31:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:35:18.381 19:31:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:35:18.381 19:31:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1209 -- # sleep 2 00:35:20.291 19:31:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:35:20.291 19:31:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:35:20.291 19:31:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:35:20.291 19:31:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:35:20.291 19:31:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:35:20.291 19:31:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1212 -- # return 0 00:35:20.291 19:31:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:35:20.291 [global] 00:35:20.291 thread=1 00:35:20.291 invalidate=1 00:35:20.292 rw=write 00:35:20.292 time_based=1 00:35:20.292 runtime=1 00:35:20.292 ioengine=libaio 00:35:20.292 direct=1 00:35:20.292 bs=4096 00:35:20.292 iodepth=1 00:35:20.292 norandommap=0 00:35:20.292 numjobs=1 00:35:20.292 00:35:20.292 verify_dump=1 00:35:20.292 verify_backlog=512 00:35:20.292 verify_state_save=0 00:35:20.292 do_verify=1 00:35:20.292 verify=crc32c-intel 00:35:20.292 [job0] 00:35:20.292 filename=/dev/nvme0n1 00:35:20.292 Could not set queue depth (nvme0n1) 00:35:20.552 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:20.552 fio-3.35 00:35:20.552 Starting 1 thread 00:35:21.936 00:35:21.936 job0: (groupid=0, jobs=1): err= 0: pid=4148180: Tue Nov 26 19:31:34 2024 00:35:21.936 read: IOPS=17, BW=71.1KiB/s (72.8kB/s)(72.0KiB/1013msec) 00:35:21.936 slat (nsec): min=25986, max=26634, avg=26166.17, stdev=160.74 00:35:21.936 clat (usec): min=40922, max=41040, avg=40966.28, stdev=29.29 00:35:21.936 lat (usec): min=40948, max=41066, avg=40992.45, stdev=29.27 00:35:21.936 clat percentiles (usec): 00:35:21.936 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:35:21.936 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:35:21.936 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:35:21.936 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:35:21.936 | 99.99th=[41157] 00:35:21.936 write: IOPS=505, BW=2022KiB/s (2070kB/s)(2048KiB/1013msec); 0 zone resets 00:35:21.936 slat (nsec): min=9375, max=65391, avg=33039.23, stdev=5944.68 00:35:21.936 clat (usec): min=208, max=758, avg=495.78, stdev=134.73 00:35:21.936 lat (usec): min=236, max=791, avg=528.82, stdev=135.51 00:35:21.936 clat percentiles (usec): 00:35:21.936 | 1.00th=[ 277], 5.00th=[ 306], 10.00th=[ 326], 20.00th=[ 355], 00:35:21.936 | 30.00th=[ 412], 40.00th=[ 429], 50.00th=[ 486], 60.00th=[ 529], 00:35:21.936 | 70.00th=[ 603], 80.00th=[ 652], 90.00th=[ 685], 95.00th=[ 709], 00:35:21.936 | 99.00th=[ 734], 99.50th=[ 742], 99.90th=[ 758], 99.95th=[ 758], 00:35:21.936 | 99.99th=[ 758] 00:35:21.936 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:35:21.936 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:21.936 lat (usec) : 250=0.75%, 500=50.38%, 750=45.28%, 1000=0.19% 00:35:21.936 lat (msec) : 50=3.40% 00:35:21.936 cpu : usr=1.58%, sys=1.78%, ctx=530, majf=0, minf=1 00:35:21.936 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:21.936 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:21.936 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:21.936 issued rwts: total=18,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:21.936 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:21.936 00:35:21.936 Run status group 0 (all jobs): 00:35:21.936 READ: bw=71.1KiB/s (72.8kB/s), 71.1KiB/s-71.1KiB/s (72.8kB/s-72.8kB/s), io=72.0KiB (73.7kB), run=1013-1013msec 00:35:21.936 WRITE: bw=2022KiB/s (2070kB/s), 2022KiB/s-2022KiB/s (2070kB/s-2070kB/s), io=2048KiB (2097kB), run=1013-1013msec 00:35:21.936 00:35:21.936 Disk stats (read/write): 00:35:21.936 nvme0n1: ios=65/512, merge=0/0, ticks=674/225, in_queue=899, util=93.69% 00:35:21.936 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:35:21.936 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:35:21.936 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:35:21.936 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1223 -- # local i=0 00:35:21.936 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:35:21.936 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:35:21.936 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:35:21.936 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1235 -- # return 0 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@516 -- # nvmfcleanup 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:22.197 rmmod nvme_tcp 00:35:22.197 rmmod nvme_fabrics 00:35:22.197 rmmod nvme_keyring 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@517 -- # '[' -n 4147048 ']' 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@518 -- # killprocess 4147048 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@954 -- # '[' -z 4147048 ']' 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@958 -- # kill -0 4147048 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@959 -- # uname 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4147048 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4147048' 00:35:22.197 killing process with pid 4147048 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@973 -- # kill 4147048 00:35:22.197 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@978 -- # wait 4147048 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-save 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-restore 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:22.458 19:31:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:24.370 19:31:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:24.370 00:35:24.370 real 0m16.777s 00:35:24.370 user 0m32.724s 00:35:24.370 sys 0m8.215s 00:35:24.370 19:31:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:24.370 19:31:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:35:24.370 ************************************ 00:35:24.370 END TEST nvmf_nmic 00:35:24.370 ************************************ 00:35:24.370 19:31:36 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:35:24.370 19:31:36 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:35:24.370 19:31:36 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:24.370 19:31:36 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:35:24.631 ************************************ 00:35:24.631 START TEST nvmf_fio_target 00:35:24.631 ************************************ 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:35:24.631 * Looking for test storage... 00:35:24.631 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1693 -- # lcov --version 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:24.631 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:35:24.632 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.632 --rc genhtml_branch_coverage=1 00:35:24.632 --rc genhtml_function_coverage=1 00:35:24.632 --rc genhtml_legend=1 00:35:24.632 --rc geninfo_all_blocks=1 00:35:24.632 --rc geninfo_unexecuted_blocks=1 00:35:24.632 00:35:24.632 ' 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:35:24.632 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.632 --rc genhtml_branch_coverage=1 00:35:24.632 --rc genhtml_function_coverage=1 00:35:24.632 --rc genhtml_legend=1 00:35:24.632 --rc geninfo_all_blocks=1 00:35:24.632 --rc geninfo_unexecuted_blocks=1 00:35:24.632 00:35:24.632 ' 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:35:24.632 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.632 --rc genhtml_branch_coverage=1 00:35:24.632 --rc genhtml_function_coverage=1 00:35:24.632 --rc genhtml_legend=1 00:35:24.632 --rc geninfo_all_blocks=1 00:35:24.632 --rc geninfo_unexecuted_blocks=1 00:35:24.632 00:35:24.632 ' 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:35:24.632 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.632 --rc genhtml_branch_coverage=1 00:35:24.632 --rc genhtml_function_coverage=1 00:35:24.632 --rc genhtml_legend=1 00:35:24.632 --rc geninfo_all_blocks=1 00:35:24.632 --rc geninfo_unexecuted_blocks=1 00:35:24.632 00:35:24.632 ' 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:35:24.632 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:24.633 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:24.893 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:35:24.893 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:35:24.893 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:35:24.893 19:31:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:35:33.043 Found 0000:31:00.0 (0x8086 - 0x159b) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:35:33.043 Found 0000:31:00.1 (0x8086 - 0x159b) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:33.043 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:35:33.044 Found net devices under 0000:31:00.0: cvl_0_0 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:35:33.044 Found net devices under 0000:31:00.1: cvl_0_1 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # is_hw=yes 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:33.044 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:33.044 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.619 ms 00:35:33.044 00:35:33.044 --- 10.0.0.2 ping statistics --- 00:35:33.044 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:33.044 rtt min/avg/max/mdev = 0.619/0.619/0.619/0.000 ms 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:33.044 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:33.044 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.241 ms 00:35:33.044 00:35:33.044 --- 10.0.0.1 ping statistics --- 00:35:33.044 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:33.044 rtt min/avg/max/mdev = 0.241/0.241/0.241/0.000 ms 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@450 -- # return 0 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@509 -- # nvmfpid=4153081 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@510 -- # waitforlisten 4153081 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@835 -- # '[' -z 4153081 ']' 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:33.044 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:33.044 19:31:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:35:33.044 [2024-11-26 19:31:45.533318] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:35:33.044 [2024-11-26 19:31:45.534812] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:35:33.044 [2024-11-26 19:31:45.534882] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:33.044 [2024-11-26 19:31:45.623469] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:35:33.044 [2024-11-26 19:31:45.660131] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:33.044 [2024-11-26 19:31:45.660166] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:33.044 [2024-11-26 19:31:45.660174] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:33.044 [2024-11-26 19:31:45.660181] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:33.044 [2024-11-26 19:31:45.660186] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:33.044 [2024-11-26 19:31:45.661689] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:33.044 [2024-11-26 19:31:45.661809] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:35:33.044 [2024-11-26 19:31:45.661949] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:33.044 [2024-11-26 19:31:45.661950] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:35:33.305 [2024-11-26 19:31:45.718426] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:35:33.305 [2024-11-26 19:31:45.718434] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:35:33.305 [2024-11-26 19:31:45.719318] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:35:33.305 [2024-11-26 19:31:45.720247] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:35:33.305 [2024-11-26 19:31:45.720312] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:35:33.877 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:33.877 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@868 -- # return 0 00:35:33.877 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:35:33.877 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:33.877 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:35:33.877 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:33.877 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:35:34.138 [2024-11-26 19:31:46.526410] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:34.138 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:34.138 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:35:34.138 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:34.400 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:35:34.400 19:31:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:34.661 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:35:34.661 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:34.661 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:35:34.661 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:35:34.923 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:35.184 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:35:35.184 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:35.184 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:35:35.184 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:35.445 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:35:35.445 19:31:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:35:35.706 19:31:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:35:35.706 19:31:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:35:35.706 19:31:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:35:35.967 19:31:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:35:35.967 19:31:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:35:36.229 19:31:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:35:36.229 [2024-11-26 19:31:48.822570] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:36.229 19:31:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:35:36.491 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:35:36.757 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:35:37.023 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:35:37.023 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1202 -- # local i=0 00:35:37.023 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:35:37.023 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1204 -- # [[ -n 4 ]] 00:35:37.023 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1205 -- # nvme_device_counter=4 00:35:37.023 19:31:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1209 -- # sleep 2 00:35:38.939 19:31:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:35:39.201 19:31:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:35:39.201 19:31:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:35:39.201 19:31:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1211 -- # nvme_devices=4 00:35:39.201 19:31:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:35:39.201 19:31:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1212 -- # return 0 00:35:39.201 19:31:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:35:39.201 [global] 00:35:39.201 thread=1 00:35:39.201 invalidate=1 00:35:39.201 rw=write 00:35:39.201 time_based=1 00:35:39.201 runtime=1 00:35:39.201 ioengine=libaio 00:35:39.201 direct=1 00:35:39.201 bs=4096 00:35:39.201 iodepth=1 00:35:39.201 norandommap=0 00:35:39.201 numjobs=1 00:35:39.201 00:35:39.201 verify_dump=1 00:35:39.201 verify_backlog=512 00:35:39.201 verify_state_save=0 00:35:39.201 do_verify=1 00:35:39.201 verify=crc32c-intel 00:35:39.201 [job0] 00:35:39.201 filename=/dev/nvme0n1 00:35:39.201 [job1] 00:35:39.201 filename=/dev/nvme0n2 00:35:39.201 [job2] 00:35:39.201 filename=/dev/nvme0n3 00:35:39.201 [job3] 00:35:39.201 filename=/dev/nvme0n4 00:35:39.201 Could not set queue depth (nvme0n1) 00:35:39.201 Could not set queue depth (nvme0n2) 00:35:39.201 Could not set queue depth (nvme0n3) 00:35:39.201 Could not set queue depth (nvme0n4) 00:35:39.462 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:39.462 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:39.462 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:39.462 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:39.462 fio-3.35 00:35:39.462 Starting 4 threads 00:35:40.849 00:35:40.849 job0: (groupid=0, jobs=1): err= 0: pid=4154510: Tue Nov 26 19:31:53 2024 00:35:40.849 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:35:40.850 slat (nsec): min=6821, max=61599, avg=26559.67, stdev=4544.66 00:35:40.850 clat (usec): min=534, max=1115, avg=891.77, stdev=83.79 00:35:40.850 lat (usec): min=561, max=1141, avg=918.33, stdev=84.53 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 644], 5.00th=[ 734], 10.00th=[ 775], 20.00th=[ 832], 00:35:40.850 | 30.00th=[ 865], 40.00th=[ 889], 50.00th=[ 906], 60.00th=[ 922], 00:35:40.850 | 70.00th=[ 938], 80.00th=[ 955], 90.00th=[ 988], 95.00th=[ 1012], 00:35:40.850 | 99.00th=[ 1045], 99.50th=[ 1074], 99.90th=[ 1123], 99.95th=[ 1123], 00:35:40.850 | 99.99th=[ 1123] 00:35:40.850 write: IOPS=923, BW=3692KiB/s (3781kB/s)(3696KiB/1001msec); 0 zone resets 00:35:40.850 slat (nsec): min=9089, max=66906, avg=29819.18, stdev=9709.49 00:35:40.850 clat (usec): min=164, max=834, avg=532.19, stdev=120.03 00:35:40.850 lat (usec): min=176, max=867, avg=562.01, stdev=124.27 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 196], 5.00th=[ 322], 10.00th=[ 379], 20.00th=[ 441], 00:35:40.850 | 30.00th=[ 478], 40.00th=[ 502], 50.00th=[ 537], 60.00th=[ 570], 00:35:40.850 | 70.00th=[ 603], 80.00th=[ 644], 90.00th=[ 693], 95.00th=[ 717], 00:35:40.850 | 99.00th=[ 766], 99.50th=[ 791], 99.90th=[ 832], 99.95th=[ 832], 00:35:40.850 | 99.99th=[ 832] 00:35:40.850 bw ( KiB/s): min= 4096, max= 4096, per=42.79%, avg=4096.00, stdev= 0.00, samples=1 00:35:40.850 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:40.850 lat (usec) : 250=1.18%, 500=24.30%, 750=40.18%, 1000=31.62% 00:35:40.850 lat (msec) : 2=2.72% 00:35:40.850 cpu : usr=3.00%, sys=5.40%, ctx=1436, majf=0, minf=2 00:35:40.850 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:40.850 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 issued rwts: total=512,924,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:40.850 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:40.850 job1: (groupid=0, jobs=1): err= 0: pid=4154511: Tue Nov 26 19:31:53 2024 00:35:40.850 read: IOPS=16, BW=66.1KiB/s (67.7kB/s)(68.0KiB/1028msec) 00:35:40.850 slat (nsec): min=25136, max=28944, avg=25588.47, stdev=895.89 00:35:40.850 clat (usec): min=1196, max=42068, avg=39517.66, stdev=9877.07 00:35:40.850 lat (usec): min=1222, max=42097, avg=39543.25, stdev=9877.03 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 1205], 5.00th=[ 1205], 10.00th=[41157], 20.00th=[41681], 00:35:40.850 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[42206], 00:35:40.850 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:35:40.850 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:35:40.850 | 99.99th=[42206] 00:35:40.850 write: IOPS=498, BW=1992KiB/s (2040kB/s)(2048KiB/1028msec); 0 zone resets 00:35:40.850 slat (nsec): min=9918, max=68392, avg=31029.01, stdev=8019.55 00:35:40.850 clat (usec): min=185, max=938, avg=656.25, stdev=121.96 00:35:40.850 lat (usec): min=198, max=971, avg=687.27, stdev=123.97 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 351], 5.00th=[ 445], 10.00th=[ 502], 20.00th=[ 553], 00:35:40.850 | 30.00th=[ 603], 40.00th=[ 635], 50.00th=[ 668], 60.00th=[ 693], 00:35:40.850 | 70.00th=[ 725], 80.00th=[ 758], 90.00th=[ 807], 95.00th=[ 840], 00:35:40.850 | 99.00th=[ 914], 99.50th=[ 922], 99.90th=[ 938], 99.95th=[ 938], 00:35:40.850 | 99.99th=[ 938] 00:35:40.850 bw ( KiB/s): min= 4096, max= 4096, per=42.79%, avg=4096.00, stdev= 0.00, samples=1 00:35:40.850 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:40.850 lat (usec) : 250=0.19%, 500=9.64%, 750=64.27%, 1000=22.68% 00:35:40.850 lat (msec) : 2=0.19%, 50=3.02% 00:35:40.850 cpu : usr=0.88%, sys=1.36%, ctx=529, majf=0, minf=1 00:35:40.850 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:40.850 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 issued rwts: total=17,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:40.850 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:40.850 job2: (groupid=0, jobs=1): err= 0: pid=4154512: Tue Nov 26 19:31:53 2024 00:35:40.850 read: IOPS=167, BW=671KiB/s (687kB/s)(684KiB/1019msec) 00:35:40.850 slat (nsec): min=6794, max=40251, avg=23748.95, stdev=5651.75 00:35:40.850 clat (usec): min=738, max=41986, avg=3884.24, stdev=10438.40 00:35:40.850 lat (usec): min=746, max=42012, avg=3907.99, stdev=10438.89 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 791], 5.00th=[ 865], 10.00th=[ 906], 20.00th=[ 963], 00:35:40.850 | 30.00th=[ 996], 40.00th=[ 1012], 50.00th=[ 1037], 60.00th=[ 1057], 00:35:40.850 | 70.00th=[ 1074], 80.00th=[ 1106], 90.00th=[ 1188], 95.00th=[41681], 00:35:40.850 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:35:40.850 | 99.99th=[42206] 00:35:40.850 write: IOPS=502, BW=2010KiB/s (2058kB/s)(2048KiB/1019msec); 0 zone resets 00:35:40.850 slat (nsec): min=9747, max=53242, avg=30848.26, stdev=8082.72 00:35:40.850 clat (usec): min=240, max=990, avg=643.81, stdev=127.00 00:35:40.850 lat (usec): min=250, max=1023, avg=674.66, stdev=129.76 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 359], 5.00th=[ 408], 10.00th=[ 465], 20.00th=[ 529], 00:35:40.850 | 30.00th=[ 586], 40.00th=[ 619], 50.00th=[ 652], 60.00th=[ 701], 00:35:40.850 | 70.00th=[ 725], 80.00th=[ 750], 90.00th=[ 791], 95.00th=[ 824], 00:35:40.850 | 99.00th=[ 906], 99.50th=[ 938], 99.90th=[ 988], 99.95th=[ 988], 00:35:40.850 | 99.99th=[ 988] 00:35:40.850 bw ( KiB/s): min= 4096, max= 4096, per=42.79%, avg=4096.00, stdev= 0.00, samples=1 00:35:40.850 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:40.850 lat (usec) : 250=0.15%, 500=10.25%, 750=48.90%, 1000=23.72% 00:35:40.850 lat (msec) : 2=15.23%, 50=1.76% 00:35:40.850 cpu : usr=0.88%, sys=2.06%, ctx=683, majf=0, minf=1 00:35:40.850 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:40.850 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 issued rwts: total=171,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:40.850 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:40.850 job3: (groupid=0, jobs=1): err= 0: pid=4154513: Tue Nov 26 19:31:53 2024 00:35:40.850 read: IOPS=18, BW=74.3KiB/s (76.1kB/s)(76.0KiB/1023msec) 00:35:40.850 slat (nsec): min=9873, max=29768, avg=24153.42, stdev=5035.01 00:35:40.850 clat (usec): min=922, max=42062, avg=34966.50, stdev=15229.61 00:35:40.850 lat (usec): min=947, max=42086, avg=34990.66, stdev=15228.59 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 922], 5.00th=[ 922], 10.00th=[ 1090], 20.00th=[32637], 00:35:40.850 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[41681], 00:35:40.850 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:35:40.850 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:35:40.850 | 99.99th=[42206] 00:35:40.850 write: IOPS=500, BW=2002KiB/s (2050kB/s)(2048KiB/1023msec); 0 zone resets 00:35:40.850 slat (nsec): min=9955, max=53359, avg=28634.50, stdev=9869.91 00:35:40.850 clat (usec): min=294, max=1062, avg=664.25, stdev=147.55 00:35:40.850 lat (usec): min=310, max=1095, avg=692.89, stdev=151.18 00:35:40.850 clat percentiles (usec): 00:35:40.850 | 1.00th=[ 347], 5.00th=[ 420], 10.00th=[ 465], 20.00th=[ 537], 00:35:40.850 | 30.00th=[ 594], 40.00th=[ 635], 50.00th=[ 676], 60.00th=[ 709], 00:35:40.850 | 70.00th=[ 742], 80.00th=[ 775], 90.00th=[ 857], 95.00th=[ 898], 00:35:40.850 | 99.00th=[ 1020], 99.50th=[ 1037], 99.90th=[ 1057], 99.95th=[ 1057], 00:35:40.850 | 99.99th=[ 1057] 00:35:40.850 bw ( KiB/s): min= 4096, max= 4096, per=42.79%, avg=4096.00, stdev= 0.00, samples=1 00:35:40.850 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:40.850 lat (usec) : 500=15.25%, 750=55.93%, 1000=24.11% 00:35:40.850 lat (msec) : 2=1.69%, 50=3.01% 00:35:40.850 cpu : usr=0.49%, sys=1.57%, ctx=531, majf=0, minf=1 00:35:40.850 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:40.850 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:40.850 issued rwts: total=19,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:40.850 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:40.850 00:35:40.850 Run status group 0 (all jobs): 00:35:40.850 READ: bw=2798KiB/s (2865kB/s), 66.1KiB/s-2046KiB/s (67.7kB/s-2095kB/s), io=2876KiB (2945kB), run=1001-1028msec 00:35:40.850 WRITE: bw=9572KiB/s (9802kB/s), 1992KiB/s-3692KiB/s (2040kB/s-3781kB/s), io=9840KiB (10.1MB), run=1001-1028msec 00:35:40.850 00:35:40.850 Disk stats (read/write): 00:35:40.850 nvme0n1: ios=562/568, merge=0/0, ticks=436/253, in_queue=689, util=86.07% 00:35:40.850 nvme0n2: ios=38/512, merge=0/0, ticks=494/327, in_queue=821, util=86.07% 00:35:40.850 nvme0n3: ios=166/512, merge=0/0, ticks=448/324, in_queue=772, util=88.21% 00:35:40.850 nvme0n4: ios=14/512, merge=0/0, ticks=455/329, in_queue=784, util=89.26% 00:35:40.850 19:31:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:35:40.850 [global] 00:35:40.850 thread=1 00:35:40.850 invalidate=1 00:35:40.850 rw=randwrite 00:35:40.850 time_based=1 00:35:40.850 runtime=1 00:35:40.850 ioengine=libaio 00:35:40.850 direct=1 00:35:40.850 bs=4096 00:35:40.850 iodepth=1 00:35:40.850 norandommap=0 00:35:40.850 numjobs=1 00:35:40.850 00:35:40.850 verify_dump=1 00:35:40.850 verify_backlog=512 00:35:40.850 verify_state_save=0 00:35:40.850 do_verify=1 00:35:40.850 verify=crc32c-intel 00:35:40.850 [job0] 00:35:40.850 filename=/dev/nvme0n1 00:35:40.850 [job1] 00:35:40.850 filename=/dev/nvme0n2 00:35:40.850 [job2] 00:35:40.850 filename=/dev/nvme0n3 00:35:40.850 [job3] 00:35:40.850 filename=/dev/nvme0n4 00:35:40.850 Could not set queue depth (nvme0n1) 00:35:40.850 Could not set queue depth (nvme0n2) 00:35:40.850 Could not set queue depth (nvme0n3) 00:35:40.850 Could not set queue depth (nvme0n4) 00:35:41.112 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:41.112 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:41.112 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:41.112 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:41.112 fio-3.35 00:35:41.112 Starting 4 threads 00:35:42.501 00:35:42.501 job0: (groupid=0, jobs=1): err= 0: pid=4155028: Tue Nov 26 19:31:54 2024 00:35:42.501 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:35:42.501 slat (nsec): min=7863, max=60024, avg=25419.67, stdev=3131.83 00:35:42.501 clat (usec): min=816, max=1351, avg=1095.25, stdev=75.87 00:35:42.501 lat (usec): min=842, max=1375, avg=1120.67, stdev=75.75 00:35:42.501 clat percentiles (usec): 00:35:42.501 | 1.00th=[ 881], 5.00th=[ 955], 10.00th=[ 1004], 20.00th=[ 1045], 00:35:42.501 | 30.00th=[ 1074], 40.00th=[ 1090], 50.00th=[ 1106], 60.00th=[ 1106], 00:35:42.501 | 70.00th=[ 1139], 80.00th=[ 1156], 90.00th=[ 1188], 95.00th=[ 1221], 00:35:42.501 | 99.00th=[ 1254], 99.50th=[ 1270], 99.90th=[ 1352], 99.95th=[ 1352], 00:35:42.501 | 99.99th=[ 1352] 00:35:42.501 write: IOPS=638, BW=2553KiB/s (2615kB/s)(2556KiB/1001msec); 0 zone resets 00:35:42.501 slat (nsec): min=9382, max=50633, avg=26319.24, stdev=9487.60 00:35:42.501 clat (usec): min=158, max=1203, avg=627.28, stdev=128.81 00:35:42.501 lat (usec): min=172, max=1234, avg=653.60, stdev=131.42 00:35:42.501 clat percentiles (usec): 00:35:42.501 | 1.00th=[ 343], 5.00th=[ 396], 10.00th=[ 465], 20.00th=[ 515], 00:35:42.501 | 30.00th=[ 570], 40.00th=[ 611], 50.00th=[ 635], 60.00th=[ 668], 00:35:42.501 | 70.00th=[ 701], 80.00th=[ 725], 90.00th=[ 783], 95.00th=[ 832], 00:35:42.501 | 99.00th=[ 922], 99.50th=[ 1004], 99.90th=[ 1205], 99.95th=[ 1205], 00:35:42.501 | 99.99th=[ 1205] 00:35:42.501 bw ( KiB/s): min= 4096, max= 4096, per=43.27%, avg=4096.00, stdev= 0.00, samples=1 00:35:42.501 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:42.501 lat (usec) : 250=0.26%, 500=9.21%, 750=37.97%, 1000=12.25% 00:35:42.501 lat (msec) : 2=40.31% 00:35:42.501 cpu : usr=1.30%, sys=3.50%, ctx=1152, majf=0, minf=1 00:35:42.501 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:42.501 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.501 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.501 issued rwts: total=512,639,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:42.501 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:42.501 job1: (groupid=0, jobs=1): err= 0: pid=4155029: Tue Nov 26 19:31:54 2024 00:35:42.501 read: IOPS=16, BW=67.5KiB/s (69.1kB/s)(68.0KiB/1007msec) 00:35:42.501 slat (nsec): min=27262, max=28306, avg=27737.59, stdev=272.90 00:35:42.501 clat (usec): min=1190, max=42061, avg=39416.54, stdev=9854.88 00:35:42.501 lat (usec): min=1218, max=42089, avg=39444.28, stdev=9854.90 00:35:42.501 clat percentiles (usec): 00:35:42.501 | 1.00th=[ 1188], 5.00th=[ 1188], 10.00th=[41157], 20.00th=[41681], 00:35:42.501 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[42206], 00:35:42.501 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:35:42.501 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:35:42.501 | 99.99th=[42206] 00:35:42.501 write: IOPS=508, BW=2034KiB/s (2083kB/s)(2048KiB/1007msec); 0 zone resets 00:35:42.501 slat (nsec): min=9242, max=71289, avg=31575.88, stdev=9692.34 00:35:42.501 clat (usec): min=316, max=1003, avg=617.36, stdev=118.92 00:35:42.501 lat (usec): min=325, max=1037, avg=648.94, stdev=121.25 00:35:42.501 clat percentiles (usec): 00:35:42.501 | 1.00th=[ 355], 5.00th=[ 424], 10.00th=[ 465], 20.00th=[ 494], 00:35:42.501 | 30.00th=[ 545], 40.00th=[ 594], 50.00th=[ 627], 60.00th=[ 660], 00:35:42.501 | 70.00th=[ 693], 80.00th=[ 725], 90.00th=[ 775], 95.00th=[ 799], 00:35:42.501 | 99.00th=[ 848], 99.50th=[ 865], 99.90th=[ 1004], 99.95th=[ 1004], 00:35:42.501 | 99.99th=[ 1004] 00:35:42.501 bw ( KiB/s): min= 4096, max= 4096, per=43.27%, avg=4096.00, stdev= 0.00, samples=1 00:35:42.501 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:42.501 lat (usec) : 500=20.60%, 750=63.33%, 1000=12.67% 00:35:42.501 lat (msec) : 2=0.38%, 50=3.02% 00:35:42.501 cpu : usr=1.09%, sys=2.09%, ctx=530, majf=0, minf=1 00:35:42.501 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:42.501 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.501 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.501 issued rwts: total=17,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:42.501 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:42.501 job2: (groupid=0, jobs=1): err= 0: pid=4155030: Tue Nov 26 19:31:54 2024 00:35:42.501 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:35:42.501 slat (nsec): min=9071, max=59994, avg=28711.88, stdev=3607.35 00:35:42.501 clat (usec): min=814, max=1355, avg=1097.29, stdev=77.31 00:35:42.501 lat (usec): min=823, max=1383, avg=1126.00, stdev=77.84 00:35:42.501 clat percentiles (usec): 00:35:42.501 | 1.00th=[ 881], 5.00th=[ 947], 10.00th=[ 996], 20.00th=[ 1045], 00:35:42.501 | 30.00th=[ 1074], 40.00th=[ 1090], 50.00th=[ 1106], 60.00th=[ 1123], 00:35:42.501 | 70.00th=[ 1139], 80.00th=[ 1156], 90.00th=[ 1188], 95.00th=[ 1221], 00:35:42.501 | 99.00th=[ 1270], 99.50th=[ 1287], 99.90th=[ 1352], 99.95th=[ 1352], 00:35:42.501 | 99.99th=[ 1352] 00:35:42.501 write: IOPS=610, BW=2442KiB/s (2500kB/s)(2444KiB/1001msec); 0 zone resets 00:35:42.501 slat (nsec): min=9267, max=67668, avg=30842.59, stdev=10571.97 00:35:42.501 clat (usec): min=290, max=1118, avg=645.68, stdev=114.54 00:35:42.501 lat (usec): min=301, max=1154, avg=676.52, stdev=120.58 00:35:42.501 clat percentiles (usec): 00:35:42.501 | 1.00th=[ 363], 5.00th=[ 433], 10.00th=[ 486], 20.00th=[ 562], 00:35:42.501 | 30.00th=[ 594], 40.00th=[ 627], 50.00th=[ 660], 60.00th=[ 685], 00:35:42.501 | 70.00th=[ 717], 80.00th=[ 742], 90.00th=[ 775], 95.00th=[ 807], 00:35:42.502 | 99.00th=[ 857], 99.50th=[ 914], 99.90th=[ 1123], 99.95th=[ 1123], 00:35:42.502 | 99.99th=[ 1123] 00:35:42.502 bw ( KiB/s): min= 4096, max= 4096, per=43.27%, avg=4096.00, stdev= 0.00, samples=1 00:35:42.502 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:42.502 lat (usec) : 500=7.03%, 750=38.11%, 1000=13.80% 00:35:42.502 lat (msec) : 2=41.05% 00:35:42.502 cpu : usr=2.30%, sys=4.50%, ctx=1125, majf=0, minf=1 00:35:42.502 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:42.502 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.502 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.502 issued rwts: total=512,611,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:42.502 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:42.502 job3: (groupid=0, jobs=1): err= 0: pid=4155031: Tue Nov 26 19:31:54 2024 00:35:42.502 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:35:42.502 slat (nsec): min=26923, max=45999, avg=27681.74, stdev=2348.60 00:35:42.502 clat (usec): min=812, max=1399, avg=1105.43, stdev=88.21 00:35:42.502 lat (usec): min=840, max=1427, avg=1133.11, stdev=88.18 00:35:42.502 clat percentiles (usec): 00:35:42.502 | 1.00th=[ 857], 5.00th=[ 947], 10.00th=[ 988], 20.00th=[ 1045], 00:35:42.502 | 30.00th=[ 1074], 40.00th=[ 1090], 50.00th=[ 1106], 60.00th=[ 1123], 00:35:42.502 | 70.00th=[ 1156], 80.00th=[ 1172], 90.00th=[ 1205], 95.00th=[ 1237], 00:35:42.502 | 99.00th=[ 1270], 99.50th=[ 1303], 99.90th=[ 1401], 99.95th=[ 1401], 00:35:42.502 | 99.99th=[ 1401] 00:35:42.502 write: IOPS=620, BW=2482KiB/s (2541kB/s)(2484KiB/1001msec); 0 zone resets 00:35:42.502 slat (nsec): min=9130, max=58457, avg=30512.70, stdev=9171.17 00:35:42.502 clat (usec): min=250, max=2364, avg=631.44, stdev=147.17 00:35:42.502 lat (usec): min=285, max=2397, avg=661.96, stdev=150.46 00:35:42.502 clat percentiles (usec): 00:35:42.502 | 1.00th=[ 306], 5.00th=[ 400], 10.00th=[ 465], 20.00th=[ 519], 00:35:42.502 | 30.00th=[ 570], 40.00th=[ 603], 50.00th=[ 635], 60.00th=[ 668], 00:35:42.502 | 70.00th=[ 709], 80.00th=[ 734], 90.00th=[ 783], 95.00th=[ 832], 00:35:42.502 | 99.00th=[ 938], 99.50th=[ 988], 99.90th=[ 2376], 99.95th=[ 2376], 00:35:42.502 | 99.99th=[ 2376] 00:35:42.502 bw ( KiB/s): min= 4096, max= 4096, per=43.27%, avg=4096.00, stdev= 0.00, samples=1 00:35:42.502 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:35:42.502 lat (usec) : 500=8.83%, 750=36.80%, 1000=14.56% 00:35:42.502 lat (msec) : 2=39.72%, 4=0.09% 00:35:42.502 cpu : usr=2.10%, sys=4.80%, ctx=1133, majf=0, minf=1 00:35:42.502 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:42.502 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.502 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:42.502 issued rwts: total=512,621,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:42.502 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:42.502 00:35:42.502 Run status group 0 (all jobs): 00:35:42.502 READ: bw=6169KiB/s (6317kB/s), 67.5KiB/s-2046KiB/s (69.1kB/s-2095kB/s), io=6212KiB (6361kB), run=1001-1007msec 00:35:42.502 WRITE: bw=9466KiB/s (9693kB/s), 2034KiB/s-2553KiB/s (2083kB/s-2615kB/s), io=9532KiB (9761kB), run=1001-1007msec 00:35:42.502 00:35:42.502 Disk stats (read/write): 00:35:42.502 nvme0n1: ios=490/512, merge=0/0, ticks=532/317, in_queue=849, util=87.68% 00:35:42.502 nvme0n2: ios=62/512, merge=0/0, ticks=851/253, in_queue=1104, util=96.74% 00:35:42.502 nvme0n3: ios=462/512, merge=0/0, ticks=1217/280, in_queue=1497, util=99.47% 00:35:42.502 nvme0n4: ios=433/512, merge=0/0, ticks=433/253, in_queue=686, util=89.52% 00:35:42.502 19:31:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:35:42.502 [global] 00:35:42.502 thread=1 00:35:42.502 invalidate=1 00:35:42.502 rw=write 00:35:42.502 time_based=1 00:35:42.502 runtime=1 00:35:42.502 ioengine=libaio 00:35:42.502 direct=1 00:35:42.502 bs=4096 00:35:42.502 iodepth=128 00:35:42.502 norandommap=0 00:35:42.502 numjobs=1 00:35:42.502 00:35:42.502 verify_dump=1 00:35:42.502 verify_backlog=512 00:35:42.502 verify_state_save=0 00:35:42.502 do_verify=1 00:35:42.502 verify=crc32c-intel 00:35:42.502 [job0] 00:35:42.502 filename=/dev/nvme0n1 00:35:42.502 [job1] 00:35:42.502 filename=/dev/nvme0n2 00:35:42.502 [job2] 00:35:42.502 filename=/dev/nvme0n3 00:35:42.502 [job3] 00:35:42.502 filename=/dev/nvme0n4 00:35:42.502 Could not set queue depth (nvme0n1) 00:35:42.502 Could not set queue depth (nvme0n2) 00:35:42.502 Could not set queue depth (nvme0n3) 00:35:42.502 Could not set queue depth (nvme0n4) 00:35:42.763 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:42.763 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:42.763 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:42.763 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:42.763 fio-3.35 00:35:42.763 Starting 4 threads 00:35:44.151 00:35:44.151 job0: (groupid=0, jobs=1): err= 0: pid=4155559: Tue Nov 26 19:31:56 2024 00:35:44.151 read: IOPS=5328, BW=20.8MiB/s (21.8MB/s)(20.9MiB/1005msec) 00:35:44.151 slat (nsec): min=937, max=8341.8k, avg=87917.25, stdev=571130.36 00:35:44.151 clat (usec): min=1132, max=29296, avg=11165.66, stdev=3391.88 00:35:44.151 lat (usec): min=5382, max=29304, avg=11253.58, stdev=3426.02 00:35:44.151 clat percentiles (usec): 00:35:44.151 | 1.00th=[ 5735], 5.00th=[ 7570], 10.00th=[ 7963], 20.00th=[ 8717], 00:35:44.151 | 30.00th=[ 9241], 40.00th=[ 9765], 50.00th=[10290], 60.00th=[10945], 00:35:44.151 | 70.00th=[11994], 80.00th=[13173], 90.00th=[15664], 95.00th=[17695], 00:35:44.151 | 99.00th=[27132], 99.50th=[27132], 99.90th=[27132], 99.95th=[29230], 00:35:44.151 | 99.99th=[29230] 00:35:44.151 write: IOPS=5603, BW=21.9MiB/s (23.0MB/s)(22.0MiB/1005msec); 0 zone resets 00:35:44.151 slat (nsec): min=1627, max=23912k, avg=88817.51, stdev=659191.93 00:35:44.151 clat (usec): min=4913, max=43499, avg=11234.52, stdev=4007.11 00:35:44.151 lat (usec): min=4917, max=43549, avg=11323.34, stdev=4079.69 00:35:44.151 clat percentiles (usec): 00:35:44.151 | 1.00th=[ 5932], 5.00th=[ 7504], 10.00th=[ 8160], 20.00th=[ 8979], 00:35:44.151 | 30.00th=[ 9110], 40.00th=[ 9372], 50.00th=[10028], 60.00th=[10552], 00:35:44.151 | 70.00th=[11994], 80.00th=[13173], 90.00th=[15401], 95.00th=[16450], 00:35:44.151 | 99.00th=[25822], 99.50th=[36963], 99.90th=[36963], 99.95th=[43254], 00:35:44.151 | 99.99th=[43254] 00:35:44.151 bw ( KiB/s): min=20464, max=24592, per=24.74%, avg=22528.00, stdev=2918.94, samples=2 00:35:44.151 iops : min= 5116, max= 6148, avg=5632.00, stdev=729.73, samples=2 00:35:44.151 lat (msec) : 2=0.01%, 10=48.39%, 20=48.65%, 50=2.95% 00:35:44.151 cpu : usr=3.69%, sys=5.78%, ctx=468, majf=0, minf=2 00:35:44.151 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:35:44.151 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:44.151 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:44.151 issued rwts: total=5355,5632,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:44.151 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:44.151 job1: (groupid=0, jobs=1): err= 0: pid=4155560: Tue Nov 26 19:31:56 2024 00:35:44.151 read: IOPS=7139, BW=27.9MiB/s (29.2MB/s)(28.0MiB/1004msec) 00:35:44.151 slat (nsec): min=940, max=11211k, avg=60321.56, stdev=506321.56 00:35:44.151 clat (usec): min=2280, max=28282, avg=8554.83, stdev=3985.35 00:35:44.151 lat (usec): min=2290, max=28308, avg=8615.15, stdev=4020.44 00:35:44.151 clat percentiles (usec): 00:35:44.151 | 1.00th=[ 2966], 5.00th=[ 4621], 10.00th=[ 5342], 20.00th=[ 5932], 00:35:44.151 | 30.00th=[ 6390], 40.00th=[ 6783], 50.00th=[ 7111], 60.00th=[ 7701], 00:35:44.151 | 70.00th=[ 9372], 80.00th=[10683], 90.00th=[13829], 95.00th=[18744], 00:35:44.151 | 99.00th=[21890], 99.50th=[22676], 99.90th=[25035], 99.95th=[25035], 00:35:44.151 | 99.99th=[28181] 00:35:44.151 write: IOPS=7486, BW=29.2MiB/s (30.7MB/s)(29.4MiB/1004msec); 0 zone resets 00:35:44.151 slat (nsec): min=1652, max=12692k, avg=63698.06, stdev=490673.02 00:35:44.151 clat (usec): min=759, max=33641, avg=8739.65, stdev=4817.41 00:35:44.151 lat (usec): min=767, max=33645, avg=8803.34, stdev=4851.57 00:35:44.151 clat percentiles (usec): 00:35:44.151 | 1.00th=[ 1516], 5.00th=[ 3097], 10.00th=[ 3949], 20.00th=[ 5080], 00:35:44.151 | 30.00th=[ 5800], 40.00th=[ 6456], 50.00th=[ 7046], 60.00th=[ 8356], 00:35:44.151 | 70.00th=[10028], 80.00th=[13042], 90.00th=[15139], 95.00th=[18482], 00:35:44.151 | 99.00th=[22676], 99.50th=[30278], 99.90th=[33424], 99.95th=[33817], 00:35:44.151 | 99.99th=[33817] 00:35:44.151 bw ( KiB/s): min=26344, max=32768, per=32.46%, avg=29556.00, stdev=4542.45, samples=2 00:35:44.151 iops : min= 6586, max= 8192, avg=7389.00, stdev=1135.61, samples=2 00:35:44.151 lat (usec) : 1000=0.14% 00:35:44.151 lat (msec) : 2=1.04%, 4=5.62%, 10=66.26%, 20=24.88%, 50=2.07% 00:35:44.151 cpu : usr=4.69%, sys=8.57%, ctx=400, majf=0, minf=1 00:35:44.151 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:35:44.151 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:44.151 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:44.151 issued rwts: total=7168,7516,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:44.151 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:44.151 job2: (groupid=0, jobs=1): err= 0: pid=4155561: Tue Nov 26 19:31:56 2024 00:35:44.151 read: IOPS=4973, BW=19.4MiB/s (20.4MB/s)(19.5MiB/1004msec) 00:35:44.151 slat (nsec): min=916, max=12509k, avg=106422.19, stdev=701518.15 00:35:44.151 clat (usec): min=1488, max=38346, avg=13869.86, stdev=5724.08 00:35:44.151 lat (usec): min=4353, max=38347, avg=13976.28, stdev=5763.61 00:35:44.151 clat percentiles (usec): 00:35:44.151 | 1.00th=[ 5276], 5.00th=[ 7439], 10.00th=[ 8225], 20.00th=[ 8848], 00:35:44.151 | 30.00th=[ 9634], 40.00th=[11076], 50.00th=[12780], 60.00th=[14091], 00:35:44.151 | 70.00th=[15533], 80.00th=[17433], 90.00th=[23200], 95.00th=[26084], 00:35:44.151 | 99.00th=[31327], 99.50th=[32900], 99.90th=[34341], 99.95th=[34341], 00:35:44.151 | 99.99th=[38536] 00:35:44.151 write: IOPS=5099, BW=19.9MiB/s (20.9MB/s)(20.0MiB/1004msec); 0 zone resets 00:35:44.151 slat (nsec): min=1577, max=8382.3k, avg=83359.89, stdev=515834.64 00:35:44.151 clat (usec): min=550, max=26981, avg=11350.94, stdev=3453.28 00:35:44.151 lat (usec): min=600, max=26986, avg=11434.30, stdev=3501.96 00:35:44.151 clat percentiles (usec): 00:35:44.151 | 1.00th=[ 3916], 5.00th=[ 5997], 10.00th=[ 8455], 20.00th=[ 8717], 00:35:44.152 | 30.00th=[ 9110], 40.00th=[10028], 50.00th=[10814], 60.00th=[11731], 00:35:44.152 | 70.00th=[12780], 80.00th=[13698], 90.00th=[15795], 95.00th=[17433], 00:35:44.152 | 99.00th=[22152], 99.50th=[23200], 99.90th=[23462], 99.95th=[24249], 00:35:44.152 | 99.99th=[26870] 00:35:44.152 bw ( KiB/s): min=20464, max=20496, per=22.49%, avg=20480.00, stdev=22.63, samples=2 00:35:44.152 iops : min= 5116, max= 5124, avg=5120.00, stdev= 5.66, samples=2 00:35:44.152 lat (usec) : 750=0.01% 00:35:44.152 lat (msec) : 2=0.23%, 4=0.30%, 10=36.59%, 20=54.96%, 50=7.92% 00:35:44.152 cpu : usr=3.89%, sys=4.89%, ctx=344, majf=0, minf=1 00:35:44.152 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:35:44.152 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:44.152 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:44.152 issued rwts: total=4993,5120,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:44.152 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:44.152 job3: (groupid=0, jobs=1): err= 0: pid=4155562: Tue Nov 26 19:31:56 2024 00:35:44.152 read: IOPS=4203, BW=16.4MiB/s (17.2MB/s)(16.5MiB/1005msec) 00:35:44.152 slat (nsec): min=922, max=19344k, avg=128630.16, stdev=796383.05 00:35:44.152 clat (usec): min=953, max=51891, avg=16327.63, stdev=8944.75 00:35:44.152 lat (usec): min=5683, max=51914, avg=16456.26, stdev=8992.09 00:35:44.152 clat percentiles (usec): 00:35:44.152 | 1.00th=[ 7635], 5.00th=[ 9110], 10.00th=[ 9503], 20.00th=[10028], 00:35:44.152 | 30.00th=[10552], 40.00th=[10945], 50.00th=[11863], 60.00th=[12780], 00:35:44.152 | 70.00th=[18744], 80.00th=[22414], 90.00th=[31065], 95.00th=[35390], 00:35:44.152 | 99.00th=[44827], 99.50th=[49021], 99.90th=[49021], 99.95th=[49021], 00:35:44.152 | 99.99th=[51643] 00:35:44.152 write: IOPS=4585, BW=17.9MiB/s (18.8MB/s)(18.0MiB/1005msec); 0 zone resets 00:35:44.152 slat (nsec): min=1589, max=8993.6k, avg=95795.02, stdev=572678.78 00:35:44.152 clat (usec): min=5850, max=37908, avg=12630.81, stdev=6071.63 00:35:44.152 lat (usec): min=6478, max=37916, avg=12726.60, stdev=6093.35 00:35:44.152 clat percentiles (usec): 00:35:44.152 | 1.00th=[ 7046], 5.00th=[ 7767], 10.00th=[ 8586], 20.00th=[ 9241], 00:35:44.152 | 30.00th=[ 9372], 40.00th=[ 9765], 50.00th=[10290], 60.00th=[10814], 00:35:44.152 | 70.00th=[11338], 80.00th=[16319], 90.00th=[20841], 95.00th=[27395], 00:35:44.152 | 99.00th=[34341], 99.50th=[38011], 99.90th=[38011], 99.95th=[38011], 00:35:44.152 | 99.99th=[38011] 00:35:44.152 bw ( KiB/s): min=17144, max=19720, per=20.24%, avg=18432.00, stdev=1821.51, samples=2 00:35:44.152 iops : min= 4286, max= 4930, avg=4608.00, stdev=455.38, samples=2 00:35:44.152 lat (usec) : 1000=0.01% 00:35:44.152 lat (msec) : 10=32.49%, 20=48.64%, 50=18.85%, 100=0.01% 00:35:44.152 cpu : usr=2.69%, sys=3.69%, ctx=343, majf=0, minf=1 00:35:44.152 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:35:44.152 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:44.152 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:44.152 issued rwts: total=4225,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:44.152 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:44.152 00:35:44.152 Run status group 0 (all jobs): 00:35:44.152 READ: bw=84.5MiB/s (88.6MB/s), 16.4MiB/s-27.9MiB/s (17.2MB/s-29.2MB/s), io=84.9MiB (89.1MB), run=1004-1005msec 00:35:44.152 WRITE: bw=88.9MiB/s (93.2MB/s), 17.9MiB/s-29.2MiB/s (18.8MB/s-30.7MB/s), io=89.4MiB (93.7MB), run=1004-1005msec 00:35:44.152 00:35:44.152 Disk stats (read/write): 00:35:44.152 nvme0n1: ios=3816/4096, merge=0/0, ticks=22758/22420, in_queue=45178, util=96.49% 00:35:44.152 nvme0n2: ios=5171/5625, merge=0/0, ticks=33759/32316, in_queue=66075, util=98.96% 00:35:44.152 nvme0n3: ios=4022/4096, merge=0/0, ticks=20699/16214, in_queue=36913, util=86.56% 00:35:44.152 nvme0n4: ios=3584/4002, merge=0/0, ticks=14103/11305, in_queue=25408, util=88.83% 00:35:44.152 19:31:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:35:44.152 [global] 00:35:44.152 thread=1 00:35:44.152 invalidate=1 00:35:44.152 rw=randwrite 00:35:44.152 time_based=1 00:35:44.152 runtime=1 00:35:44.152 ioengine=libaio 00:35:44.152 direct=1 00:35:44.152 bs=4096 00:35:44.152 iodepth=128 00:35:44.152 norandommap=0 00:35:44.152 numjobs=1 00:35:44.152 00:35:44.152 verify_dump=1 00:35:44.152 verify_backlog=512 00:35:44.152 verify_state_save=0 00:35:44.152 do_verify=1 00:35:44.152 verify=crc32c-intel 00:35:44.152 [job0] 00:35:44.152 filename=/dev/nvme0n1 00:35:44.152 [job1] 00:35:44.152 filename=/dev/nvme0n2 00:35:44.152 [job2] 00:35:44.152 filename=/dev/nvme0n3 00:35:44.152 [job3] 00:35:44.152 filename=/dev/nvme0n4 00:35:44.152 Could not set queue depth (nvme0n1) 00:35:44.152 Could not set queue depth (nvme0n2) 00:35:44.152 Could not set queue depth (nvme0n3) 00:35:44.152 Could not set queue depth (nvme0n4) 00:35:44.722 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:44.722 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:44.722 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:44.722 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:35:44.722 fio-3.35 00:35:44.722 Starting 4 threads 00:35:45.664 00:35:45.664 job0: (groupid=0, jobs=1): err= 0: pid=4156081: Tue Nov 26 19:31:58 2024 00:35:45.664 read: IOPS=5481, BW=21.4MiB/s (22.5MB/s)(21.5MiB/1003msec) 00:35:45.664 slat (nsec): min=976, max=17753k, avg=90009.73, stdev=760730.31 00:35:45.664 clat (usec): min=1583, max=43851, avg=12274.19, stdev=6041.05 00:35:45.664 lat (usec): min=2383, max=43857, avg=12364.20, stdev=6102.16 00:35:45.664 clat percentiles (usec): 00:35:45.664 | 1.00th=[ 4490], 5.00th=[ 6128], 10.00th=[ 6325], 20.00th=[ 6783], 00:35:45.665 | 30.00th=[ 7635], 40.00th=[ 8848], 50.00th=[10945], 60.00th=[12911], 00:35:45.665 | 70.00th=[14222], 80.00th=[16909], 90.00th=[20579], 95.00th=[24249], 00:35:45.665 | 99.00th=[27919], 99.50th=[33817], 99.90th=[34341], 99.95th=[43779], 00:35:45.665 | 99.99th=[43779] 00:35:45.665 write: IOPS=5615, BW=21.9MiB/s (23.0MB/s)(22.0MiB/1003msec); 0 zone resets 00:35:45.665 slat (nsec): min=1632, max=17236k, avg=83203.40, stdev=652065.82 00:35:45.665 clat (usec): min=637, max=49733, avg=10574.78, stdev=6162.04 00:35:45.665 lat (usec): min=650, max=49742, avg=10657.98, stdev=6208.45 00:35:45.665 clat percentiles (usec): 00:35:45.665 | 1.00th=[ 2966], 5.00th=[ 3949], 10.00th=[ 5735], 20.00th=[ 6849], 00:35:45.665 | 30.00th=[ 7439], 40.00th=[ 8225], 50.00th=[ 8979], 60.00th=[10552], 00:35:45.665 | 70.00th=[11600], 80.00th=[13960], 90.00th=[15795], 95.00th=[20055], 00:35:45.665 | 99.00th=[39584], 99.50th=[46924], 99.90th=[49021], 99.95th=[49546], 00:35:45.665 | 99.99th=[49546] 00:35:45.665 bw ( KiB/s): min=22336, max=22720, per=23.00%, avg=22528.00, stdev=271.53, samples=2 00:35:45.665 iops : min= 5584, max= 5680, avg=5632.00, stdev=67.88, samples=2 00:35:45.665 lat (usec) : 750=0.03%, 1000=0.04% 00:35:45.665 lat (msec) : 2=0.16%, 4=2.58%, 10=49.40%, 20=38.92%, 50=8.88% 00:35:45.665 cpu : usr=4.09%, sys=6.09%, ctx=339, majf=0, minf=2 00:35:45.665 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:35:45.665 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:45.665 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:45.665 issued rwts: total=5498,5632,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:45.665 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:45.665 job1: (groupid=0, jobs=1): err= 0: pid=4156082: Tue Nov 26 19:31:58 2024 00:35:45.665 read: IOPS=6667, BW=26.0MiB/s (27.3MB/s)(26.3MiB/1010msec) 00:35:45.665 slat (nsec): min=889, max=21644k, avg=74724.88, stdev=537028.72 00:35:45.665 clat (usec): min=2087, max=52114, avg=9241.09, stdev=4029.25 00:35:45.665 lat (usec): min=2119, max=52123, avg=9315.81, stdev=4076.52 00:35:45.665 clat percentiles (usec): 00:35:45.665 | 1.00th=[ 4080], 5.00th=[ 6063], 10.00th=[ 6718], 20.00th=[ 7111], 00:35:45.665 | 30.00th=[ 7373], 40.00th=[ 7832], 50.00th=[ 8586], 60.00th=[ 9110], 00:35:45.665 | 70.00th=[ 9372], 80.00th=[10159], 90.00th=[11994], 95.00th=[16188], 00:35:45.665 | 99.00th=[23987], 99.50th=[30802], 99.90th=[47973], 99.95th=[52167], 00:35:45.665 | 99.99th=[52167] 00:35:45.665 write: IOPS=7097, BW=27.7MiB/s (29.1MB/s)(28.0MiB/1010msec); 0 zone resets 00:35:45.665 slat (nsec): min=1470, max=7919.4k, avg=61573.76, stdev=387802.03 00:35:45.665 clat (usec): min=1052, max=79045, avg=9188.89, stdev=8395.99 00:35:45.665 lat (usec): min=1060, max=79053, avg=9250.47, stdev=8442.40 00:35:45.665 clat percentiles (usec): 00:35:45.665 | 1.00th=[ 3359], 5.00th=[ 4490], 10.00th=[ 6259], 20.00th=[ 6915], 00:35:45.665 | 30.00th=[ 7177], 40.00th=[ 7308], 50.00th=[ 7439], 60.00th=[ 7832], 00:35:45.665 | 70.00th=[ 8356], 80.00th=[ 8848], 90.00th=[11076], 95.00th=[15139], 00:35:45.665 | 99.00th=[65799], 99.50th=[72877], 99.90th=[74974], 99.95th=[79168], 00:35:45.665 | 99.99th=[79168] 00:35:45.665 bw ( KiB/s): min=24952, max=31992, per=29.06%, avg=28472.00, stdev=4978.03, samples=2 00:35:45.665 iops : min= 6238, max= 7998, avg=7118.00, stdev=1244.51, samples=2 00:35:45.665 lat (msec) : 2=0.04%, 4=2.30%, 10=80.98%, 20=13.75%, 50=2.01% 00:35:45.665 lat (msec) : 100=0.91% 00:35:45.665 cpu : usr=3.17%, sys=7.04%, ctx=584, majf=0, minf=1 00:35:45.665 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.5% 00:35:45.665 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:45.665 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:45.665 issued rwts: total=6734,7168,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:45.665 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:45.665 job2: (groupid=0, jobs=1): err= 0: pid=4156083: Tue Nov 26 19:31:58 2024 00:35:45.665 read: IOPS=6629, BW=25.9MiB/s (27.2MB/s)(26.0MiB/1004msec) 00:35:45.665 slat (nsec): min=926, max=10556k, avg=78065.83, stdev=489215.65 00:35:45.665 clat (usec): min=3890, max=33780, avg=10008.57, stdev=3227.17 00:35:45.665 lat (usec): min=3897, max=33797, avg=10086.64, stdev=3247.80 00:35:45.665 clat percentiles (usec): 00:35:45.665 | 1.00th=[ 5342], 5.00th=[ 6325], 10.00th=[ 7177], 20.00th=[ 8094], 00:35:45.665 | 30.00th=[ 8586], 40.00th=[ 9110], 50.00th=[ 9634], 60.00th=[10028], 00:35:45.665 | 70.00th=[10421], 80.00th=[11338], 90.00th=[12780], 95.00th=[13829], 00:35:45.665 | 99.00th=[23987], 99.50th=[28967], 99.90th=[32113], 99.95th=[32113], 00:35:45.665 | 99.99th=[33817] 00:35:45.665 write: IOPS=6787, BW=26.5MiB/s (27.8MB/s)(26.6MiB/1004msec); 0 zone resets 00:35:45.665 slat (nsec): min=1509, max=6080.4k, avg=65601.09, stdev=365601.92 00:35:45.665 clat (usec): min=1224, max=18009, avg=8912.74, stdev=2038.66 00:35:45.665 lat (usec): min=1233, max=18017, avg=8978.34, stdev=2044.79 00:35:45.665 clat percentiles (usec): 00:35:45.665 | 1.00th=[ 3097], 5.00th=[ 5604], 10.00th=[ 6718], 20.00th=[ 7570], 00:35:45.665 | 30.00th=[ 7767], 40.00th=[ 8225], 50.00th=[ 8848], 60.00th=[ 9765], 00:35:45.665 | 70.00th=[10028], 80.00th=[10421], 90.00th=[11207], 95.00th=[12256], 00:35:45.665 | 99.00th=[13829], 99.50th=[14091], 99.90th=[16188], 99.95th=[17957], 00:35:45.665 | 99.99th=[17957] 00:35:45.665 bw ( KiB/s): min=24832, max=28672, per=27.31%, avg=26752.00, stdev=2715.29, samples=2 00:35:45.665 iops : min= 6208, max= 7168, avg=6688.00, stdev=678.82, samples=2 00:35:45.665 lat (msec) : 2=0.19%, 4=0.82%, 10=64.88%, 20=32.86%, 50=1.25% 00:35:45.665 cpu : usr=2.69%, sys=6.88%, ctx=692, majf=0, minf=1 00:35:45.665 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.5% 00:35:45.665 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:45.665 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:45.665 issued rwts: total=6656,6815,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:45.665 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:45.665 job3: (groupid=0, jobs=1): err= 0: pid=4156084: Tue Nov 26 19:31:58 2024 00:35:45.665 read: IOPS=5082, BW=19.9MiB/s (20.8MB/s)(20.0MiB/1006msec) 00:35:45.665 slat (nsec): min=944, max=16622k, avg=92828.46, stdev=840320.42 00:35:45.665 clat (usec): min=2190, max=36205, avg=12896.30, stdev=5371.55 00:35:45.665 lat (usec): min=2227, max=36233, avg=12989.13, stdev=5427.84 00:35:45.665 clat percentiles (usec): 00:35:45.665 | 1.00th=[ 3195], 5.00th=[ 6063], 10.00th=[ 7635], 20.00th=[ 8455], 00:35:45.665 | 30.00th=[ 9110], 40.00th=[10290], 50.00th=[11338], 60.00th=[13042], 00:35:45.665 | 70.00th=[15795], 80.00th=[17695], 90.00th=[21365], 95.00th=[22676], 00:35:45.665 | 99.00th=[26870], 99.50th=[31851], 99.90th=[34341], 99.95th=[34341], 00:35:45.665 | 99.99th=[36439] 00:35:45.665 write: IOPS=5089, BW=19.9MiB/s (20.8MB/s)(20.0MiB/1006msec); 0 zone resets 00:35:45.665 slat (nsec): min=1537, max=23782k, avg=94233.38, stdev=760975.37 00:35:45.665 clat (usec): min=1182, max=33694, avg=12040.25, stdev=5896.82 00:35:45.665 lat (usec): min=1193, max=33724, avg=12134.48, stdev=5934.44 00:35:45.665 clat percentiles (usec): 00:35:45.665 | 1.00th=[ 2212], 5.00th=[ 5014], 10.00th=[ 6718], 20.00th=[ 7898], 00:35:45.665 | 30.00th=[ 8717], 40.00th=[ 9765], 50.00th=[10683], 60.00th=[11863], 00:35:45.665 | 70.00th=[12911], 80.00th=[15139], 90.00th=[21103], 95.00th=[25560], 00:35:45.665 | 99.00th=[31065], 99.50th=[31851], 99.90th=[31851], 99.95th=[31851], 00:35:45.665 | 99.99th=[33817] 00:35:45.665 bw ( KiB/s): min=19784, max=21176, per=20.91%, avg=20480.00, stdev=984.29, samples=2 00:35:45.665 iops : min= 4946, max= 5294, avg=5120.00, stdev=246.07, samples=2 00:35:45.665 lat (msec) : 2=0.34%, 4=2.28%, 10=37.28%, 20=49.29%, 50=10.81% 00:35:45.665 cpu : usr=3.48%, sys=5.67%, ctx=261, majf=0, minf=1 00:35:45.665 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:35:45.665 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:45.665 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:35:45.665 issued rwts: total=5113,5120,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:45.665 latency : target=0, window=0, percentile=100.00%, depth=128 00:35:45.665 00:35:45.665 Run status group 0 (all jobs): 00:35:45.665 READ: bw=92.8MiB/s (97.3MB/s), 19.9MiB/s-26.0MiB/s (20.8MB/s-27.3MB/s), io=93.8MiB (98.3MB), run=1003-1010msec 00:35:45.665 WRITE: bw=95.7MiB/s (100MB/s), 19.9MiB/s-27.7MiB/s (20.8MB/s-29.1MB/s), io=96.6MiB (101MB), run=1003-1010msec 00:35:45.665 00:35:45.665 Disk stats (read/write): 00:35:45.665 nvme0n1: ios=4147/4118, merge=0/0, ticks=54065/46720, in_queue=100785, util=91.48% 00:35:45.665 nvme0n2: ios=6185/6542, merge=0/0, ticks=27285/24188, in_queue=51473, util=98.06% 00:35:45.665 nvme0n3: ios=5578/5632, merge=0/0, ticks=25117/23527, in_queue=48644, util=91.34% 00:35:45.665 nvme0n4: ios=4153/4261, merge=0/0, ticks=51482/45184, in_queue=96666, util=95.19% 00:35:45.665 19:31:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:35:45.926 19:31:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=4156344 00:35:45.926 19:31:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:35:45.926 19:31:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:35:45.926 [global] 00:35:45.926 thread=1 00:35:45.926 invalidate=1 00:35:45.926 rw=read 00:35:45.926 time_based=1 00:35:45.926 runtime=10 00:35:45.926 ioengine=libaio 00:35:45.926 direct=1 00:35:45.926 bs=4096 00:35:45.926 iodepth=1 00:35:45.926 norandommap=1 00:35:45.926 numjobs=1 00:35:45.926 00:35:45.926 [job0] 00:35:45.926 filename=/dev/nvme0n1 00:35:45.926 [job1] 00:35:45.926 filename=/dev/nvme0n2 00:35:45.926 [job2] 00:35:45.926 filename=/dev/nvme0n3 00:35:45.926 [job3] 00:35:45.926 filename=/dev/nvme0n4 00:35:45.926 Could not set queue depth (nvme0n1) 00:35:45.926 Could not set queue depth (nvme0n2) 00:35:45.926 Could not set queue depth (nvme0n3) 00:35:45.926 Could not set queue depth (nvme0n4) 00:35:46.187 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:46.187 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:46.187 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:46.187 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:35:46.187 fio-3.35 00:35:46.187 Starting 4 threads 00:35:48.732 19:32:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:35:48.993 19:32:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:35:48.993 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=901120, buflen=4096 00:35:48.993 fio: pid=4156606, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:35:49.254 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=274432, buflen=4096 00:35:49.254 fio: pid=4156605, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:35:49.254 19:32:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:35:49.254 19:32:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:35:49.254 fio: io_u error on file /dev/nvme0n1: Operation not supported: read offset=4124672, buflen=4096 00:35:49.254 fio: pid=4156603, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:35:49.254 19:32:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:35:49.255 19:32:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:35:49.517 fio: io_u error on file /dev/nvme0n2: Input/output error: read offset=483328, buflen=4096 00:35:49.517 fio: pid=4156604, err=5/file:io_u.c:1889, func=io_u error, error=Input/output error 00:35:49.517 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:35:49.517 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:35:49.517 00:35:49.517 job0: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=4156603: Tue Nov 26 19:32:02 2024 00:35:49.517 read: IOPS=338, BW=1351KiB/s (1384kB/s)(4028KiB/2981msec) 00:35:49.517 slat (usec): min=5, max=31157, avg=68.96, stdev=1026.20 00:35:49.517 clat (usec): min=300, max=42425, avg=2862.00, stdev=8336.71 00:35:49.517 lat (usec): min=326, max=42451, avg=2931.01, stdev=8389.75 00:35:49.517 clat percentiles (usec): 00:35:49.517 | 1.00th=[ 611], 5.00th=[ 701], 10.00th=[ 766], 20.00th=[ 848], 00:35:49.517 | 30.00th=[ 1029], 40.00th=[ 1106], 50.00th=[ 1123], 60.00th=[ 1139], 00:35:49.517 | 70.00th=[ 1156], 80.00th=[ 1188], 90.00th=[ 1221], 95.00th=[ 1303], 00:35:49.517 | 99.00th=[41157], 99.50th=[41681], 99.90th=[42206], 99.95th=[42206], 00:35:49.517 | 99.99th=[42206] 00:35:49.517 bw ( KiB/s): min= 960, max= 1408, per=70.76%, avg=1264.00, stdev=186.85, samples=5 00:35:49.517 iops : min= 240, max= 352, avg=316.00, stdev=46.71, samples=5 00:35:49.517 lat (usec) : 500=0.30%, 750=7.94%, 1000=20.24% 00:35:49.517 lat (msec) : 2=66.77%, 4=0.10%, 50=4.56% 00:35:49.517 cpu : usr=0.27%, sys=1.01%, ctx=1012, majf=0, minf=1 00:35:49.517 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:49.517 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.517 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.517 issued rwts: total=1008,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:49.517 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:49.517 job1: (groupid=0, jobs=1): err= 5 (file:io_u.c:1889, func=io_u error, error=Input/output error): pid=4156604: Tue Nov 26 19:32:02 2024 00:35:49.517 read: IOPS=37, BW=149KiB/s (153kB/s)(472KiB/3162msec) 00:35:49.517 slat (usec): min=6, max=9030, avg=238.86, stdev=1200.47 00:35:49.517 clat (usec): min=492, max=42087, avg=26543.98, stdev=19931.31 00:35:49.517 lat (usec): min=501, max=46930, avg=26724.56, stdev=19855.06 00:35:49.517 clat percentiles (usec): 00:35:49.517 | 1.00th=[ 515], 5.00th=[ 627], 10.00th=[ 742], 20.00th=[ 824], 00:35:49.517 | 30.00th=[ 947], 40.00th=[41157], 50.00th=[41681], 60.00th=[41681], 00:35:49.517 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:35:49.517 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:35:49.517 | 99.99th=[42206] 00:35:49.517 bw ( KiB/s): min= 96, max= 384, per=8.06%, avg=144.00, stdev=117.58, samples=6 00:35:49.517 iops : min= 24, max= 96, avg=36.00, stdev=29.39, samples=6 00:35:49.517 lat (usec) : 500=0.84%, 750=10.08%, 1000=24.37% 00:35:49.517 lat (msec) : 2=1.68%, 50=62.18% 00:35:49.517 cpu : usr=0.13%, sys=0.22%, ctx=124, majf=0, minf=2 00:35:49.518 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:49.518 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.518 complete : 0=0.8%, 4=99.2%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.518 issued rwts: total=119,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:49.518 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:49.518 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=4156605: Tue Nov 26 19:32:02 2024 00:35:49.518 read: IOPS=24, BW=96.1KiB/s (98.4kB/s)(268KiB/2788msec) 00:35:49.518 slat (usec): min=8, max=256, avg=29.36, stdev=28.10 00:35:49.518 clat (usec): min=568, max=42056, avg=41236.83, stdev=5053.08 00:35:49.518 lat (usec): min=605, max=42083, avg=41266.24, stdev=5052.25 00:35:49.518 clat percentiles (usec): 00:35:49.518 | 1.00th=[ 570], 5.00th=[41157], 10.00th=[41157], 20.00th=[41681], 00:35:49.518 | 30.00th=[41681], 40.00th=[41681], 50.00th=[42206], 60.00th=[42206], 00:35:49.518 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:35:49.518 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:35:49.518 | 99.99th=[42206] 00:35:49.518 bw ( KiB/s): min= 96, max= 96, per=5.37%, avg=96.00, stdev= 0.00, samples=5 00:35:49.518 iops : min= 24, max= 24, avg=24.00, stdev= 0.00, samples=5 00:35:49.518 lat (usec) : 750=1.47% 00:35:49.518 lat (msec) : 50=97.06% 00:35:49.518 cpu : usr=0.11%, sys=0.00%, ctx=69, majf=0, minf=2 00:35:49.518 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:49.518 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.518 complete : 0=1.4%, 4=98.6%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.518 issued rwts: total=68,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:49.518 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:49.518 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=4156606: Tue Nov 26 19:32:02 2024 00:35:49.518 read: IOPS=84, BW=336KiB/s (344kB/s)(880KiB/2619msec) 00:35:49.518 slat (nsec): min=5401, max=40164, avg=19325.19, stdev=9430.86 00:35:49.518 clat (usec): min=419, max=41944, avg=11777.33, stdev=17948.98 00:35:49.518 lat (usec): min=445, max=41971, avg=11796.62, stdev=17953.28 00:35:49.518 clat percentiles (usec): 00:35:49.518 | 1.00th=[ 515], 5.00th=[ 668], 10.00th=[ 717], 20.00th=[ 766], 00:35:49.518 | 30.00th=[ 783], 40.00th=[ 832], 50.00th=[ 865], 60.00th=[ 898], 00:35:49.518 | 70.00th=[ 971], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:35:49.518 | 99.00th=[41681], 99.50th=[41681], 99.90th=[42206], 99.95th=[42206], 00:35:49.518 | 99.99th=[42206] 00:35:49.518 bw ( KiB/s): min= 96, max= 1336, per=19.43%, avg=347.20, stdev=552.77, samples=5 00:35:49.518 iops : min= 24, max= 334, avg=86.80, stdev=138.19, samples=5 00:35:49.518 lat (usec) : 500=0.90%, 750=16.29%, 1000=54.75% 00:35:49.518 lat (msec) : 2=0.45%, 50=27.15% 00:35:49.518 cpu : usr=0.00%, sys=0.27%, ctx=221, majf=0, minf=2 00:35:49.518 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:49.518 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.518 complete : 0=0.5%, 4=99.5%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:49.518 issued rwts: total=221,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:49.518 latency : target=0, window=0, percentile=100.00%, depth=1 00:35:49.518 00:35:49.518 Run status group 0 (all jobs): 00:35:49.518 READ: bw=1786KiB/s (1829kB/s), 96.1KiB/s-1351KiB/s (98.4kB/s-1384kB/s), io=5648KiB (5784kB), run=2619-3162msec 00:35:49.518 00:35:49.518 Disk stats (read/write): 00:35:49.518 nvme0n1: ios=975/0, merge=0/0, ticks=2738/0, in_queue=2738, util=93.39% 00:35:49.518 nvme0n2: ios=116/0, merge=0/0, ticks=3048/0, in_queue=3048, util=95.17% 00:35:49.518 nvme0n3: ios=62/0, merge=0/0, ticks=2555/0, in_queue=2555, util=95.99% 00:35:49.518 nvme0n4: ios=219/0, merge=0/0, ticks=2554/0, in_queue=2554, util=96.39% 00:35:49.779 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:35:49.779 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:35:50.041 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:35:50.041 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:35:50.041 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:35:50.041 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:35:50.302 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:35:50.302 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:35:50.609 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:35:50.609 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # wait 4156344 00:35:50.609 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:35:50.609 19:32:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:35:50.609 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1223 -- # local i=0 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1235 -- # return 0 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:35:50.609 nvmf hotplug test: fio failed as expected 00:35:50.609 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:50.947 rmmod nvme_tcp 00:35:50.947 rmmod nvme_fabrics 00:35:50.947 rmmod nvme_keyring 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@517 -- # '[' -n 4153081 ']' 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@518 -- # killprocess 4153081 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@954 -- # '[' -z 4153081 ']' 00:35:50.947 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@958 -- # kill -0 4153081 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@959 -- # uname 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4153081 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4153081' 00:35:50.948 killing process with pid 4153081 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@973 -- # kill 4153081 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@978 -- # wait 4153081 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-save 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-restore 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:50.948 19:32:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:53.578 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:53.579 00:35:53.579 real 0m28.598s 00:35:53.579 user 2m12.756s 00:35:53.579 sys 0m12.745s 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:35:53.579 ************************************ 00:35:53.579 END TEST nvmf_fio_target 00:35:53.579 ************************************ 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:35:53.579 ************************************ 00:35:53.579 START TEST nvmf_bdevio 00:35:53.579 ************************************ 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:35:53.579 * Looking for test storage... 00:35:53.579 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1693 -- # lcov --version 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:35:53.579 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:53.579 --rc genhtml_branch_coverage=1 00:35:53.579 --rc genhtml_function_coverage=1 00:35:53.579 --rc genhtml_legend=1 00:35:53.579 --rc geninfo_all_blocks=1 00:35:53.579 --rc geninfo_unexecuted_blocks=1 00:35:53.579 00:35:53.579 ' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:35:53.579 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:53.579 --rc genhtml_branch_coverage=1 00:35:53.579 --rc genhtml_function_coverage=1 00:35:53.579 --rc genhtml_legend=1 00:35:53.579 --rc geninfo_all_blocks=1 00:35:53.579 --rc geninfo_unexecuted_blocks=1 00:35:53.579 00:35:53.579 ' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:35:53.579 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:53.579 --rc genhtml_branch_coverage=1 00:35:53.579 --rc genhtml_function_coverage=1 00:35:53.579 --rc genhtml_legend=1 00:35:53.579 --rc geninfo_all_blocks=1 00:35:53.579 --rc geninfo_unexecuted_blocks=1 00:35:53.579 00:35:53.579 ' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:35:53.579 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:53.579 --rc genhtml_branch_coverage=1 00:35:53.579 --rc genhtml_function_coverage=1 00:35:53.579 --rc genhtml_legend=1 00:35:53.579 --rc geninfo_all_blocks=1 00:35:53.579 --rc geninfo_unexecuted_blocks=1 00:35:53.579 00:35:53.579 ' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@476 -- # prepare_net_devs 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@438 -- # local -g is_hw=no 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # remove_spdk_ns 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:35:53.579 19:32:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:36:01.719 Found 0000:31:00.0 (0x8086 - 0x159b) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:36:01.719 Found 0000:31:00.1 (0x8086 - 0x159b) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:36:01.719 Found net devices under 0000:31:00.0: cvl_0_0 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:36:01.719 Found net devices under 0000:31:00.1: cvl_0_1 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:01.719 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # is_hw=yes 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:01.720 19:32:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:01.720 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:01.720 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.643 ms 00:36:01.720 00:36:01.720 --- 10.0.0.2 ping statistics --- 00:36:01.720 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:01.720 rtt min/avg/max/mdev = 0.643/0.643/0.643/0.000 ms 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:01.720 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:01.720 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.312 ms 00:36:01.720 00:36:01.720 --- 10.0.0.1 ping statistics --- 00:36:01.720 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:01.720 rtt min/avg/max/mdev = 0.312/0.312/0.312/0.000 ms 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@450 -- # return 0 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@509 -- # nvmfpid=4162065 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@510 -- # waitforlisten 4162065 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x78 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@835 -- # '[' -z 4162065 ']' 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:01.720 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:01.720 19:32:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:01.720 [2024-11-26 19:32:14.232755] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:36:01.720 [2024-11-26 19:32:14.233806] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:36:01.720 [2024-11-26 19:32:14.233852] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:01.720 [2024-11-26 19:32:14.339911] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:36:01.982 [2024-11-26 19:32:14.390265] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:01.982 [2024-11-26 19:32:14.390316] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:01.982 [2024-11-26 19:32:14.390325] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:01.982 [2024-11-26 19:32:14.390332] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:01.982 [2024-11-26 19:32:14.390338] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:01.982 [2024-11-26 19:32:14.392726] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:36:01.982 [2024-11-26 19:32:14.392907] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:36:01.982 [2024-11-26 19:32:14.393085] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:36:01.982 [2024-11-26 19:32:14.393086] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:36:01.982 [2024-11-26 19:32:14.479589] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:36:01.982 [2024-11-26 19:32:14.480611] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:36:01.982 [2024-11-26 19:32:14.480772] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:36:01.982 [2024-11-26 19:32:14.481404] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:36:01.982 [2024-11-26 19:32:14.481447] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@868 -- # return 0 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:02.553 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:02.553 [2024-11-26 19:32:15.078056] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:02.554 Malloc0 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:02.554 [2024-11-26 19:32:15.150273] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@560 -- # config=() 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@560 -- # local subsystem config 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:36:02.554 { 00:36:02.554 "params": { 00:36:02.554 "name": "Nvme$subsystem", 00:36:02.554 "trtype": "$TEST_TRANSPORT", 00:36:02.554 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:02.554 "adrfam": "ipv4", 00:36:02.554 "trsvcid": "$NVMF_PORT", 00:36:02.554 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:02.554 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:02.554 "hdgst": ${hdgst:-false}, 00:36:02.554 "ddgst": ${ddgst:-false} 00:36:02.554 }, 00:36:02.554 "method": "bdev_nvme_attach_controller" 00:36:02.554 } 00:36:02.554 EOF 00:36:02.554 )") 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@582 -- # cat 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@584 -- # jq . 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@585 -- # IFS=, 00:36:02.554 19:32:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:36:02.554 "params": { 00:36:02.554 "name": "Nvme1", 00:36:02.554 "trtype": "tcp", 00:36:02.554 "traddr": "10.0.0.2", 00:36:02.554 "adrfam": "ipv4", 00:36:02.554 "trsvcid": "4420", 00:36:02.554 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:36:02.554 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:36:02.554 "hdgst": false, 00:36:02.554 "ddgst": false 00:36:02.554 }, 00:36:02.554 "method": "bdev_nvme_attach_controller" 00:36:02.554 }' 00:36:02.815 [2024-11-26 19:32:15.207492] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:36:02.815 [2024-11-26 19:32:15.207549] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4162356 ] 00:36:02.815 [2024-11-26 19:32:15.286650] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:36:02.815 [2024-11-26 19:32:15.325497] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:02.815 [2024-11-26 19:32:15.325612] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:36:02.815 [2024-11-26 19:32:15.325615] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:03.076 I/O targets: 00:36:03.076 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:36:03.076 00:36:03.076 00:36:03.076 CUnit - A unit testing framework for C - Version 2.1-3 00:36:03.076 http://cunit.sourceforge.net/ 00:36:03.076 00:36:03.076 00:36:03.076 Suite: bdevio tests on: Nvme1n1 00:36:03.076 Test: blockdev write read block ...passed 00:36:03.076 Test: blockdev write zeroes read block ...passed 00:36:03.076 Test: blockdev write zeroes read no split ...passed 00:36:03.076 Test: blockdev write zeroes read split ...passed 00:36:03.076 Test: blockdev write zeroes read split partial ...passed 00:36:03.076 Test: blockdev reset ...[2024-11-26 19:32:15.619194] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:36:03.076 [2024-11-26 19:32:15.619262] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x23d74b0 (9): Bad file descriptor 00:36:03.076 [2024-11-26 19:32:15.667398] bdev_nvme.c:2282:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:36:03.076 passed 00:36:03.076 Test: blockdev write read 8 blocks ...passed 00:36:03.337 Test: blockdev write read size > 128k ...passed 00:36:03.337 Test: blockdev write read invalid size ...passed 00:36:03.337 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:36:03.337 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:36:03.337 Test: blockdev write read max offset ...passed 00:36:03.337 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:36:03.337 Test: blockdev writev readv 8 blocks ...passed 00:36:03.337 Test: blockdev writev readv 30 x 1block ...passed 00:36:03.337 Test: blockdev writev readv block ...passed 00:36:03.337 Test: blockdev writev readv size > 128k ...passed 00:36:03.337 Test: blockdev writev readv size > 128k in two iovs ...passed 00:36:03.337 Test: blockdev comparev and writev ...[2024-11-26 19:32:15.928994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.929018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:36:03.337 [2024-11-26 19:32:15.929029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.929036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:36:03.337 [2024-11-26 19:32:15.929424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.929432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:36:03.337 [2024-11-26 19:32:15.929443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.929452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:36:03.337 [2024-11-26 19:32:15.929879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.929887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:36:03.337 [2024-11-26 19:32:15.929897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.929903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:36:03.337 [2024-11-26 19:32:15.930288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.930295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:36:03.337 [2024-11-26 19:32:15.930305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:36:03.337 [2024-11-26 19:32:15.930310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:36:03.599 passed 00:36:03.599 Test: blockdev nvme passthru rw ...passed 00:36:03.599 Test: blockdev nvme passthru vendor specific ...[2024-11-26 19:32:16.013309] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:36:03.599 [2024-11-26 19:32:16.013320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:03.599 [2024-11-26 19:32:16.013570] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:36:03.599 [2024-11-26 19:32:16.013577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:36:03.599 [2024-11-26 19:32:16.013793] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:36:03.599 [2024-11-26 19:32:16.013801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:36:03.599 [2024-11-26 19:32:16.014037] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:36:03.599 [2024-11-26 19:32:16.014044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:36:03.599 passed 00:36:03.599 Test: blockdev nvme admin passthru ...passed 00:36:03.599 Test: blockdev copy ...passed 00:36:03.599 00:36:03.599 Run Summary: Type Total Ran Passed Failed Inactive 00:36:03.599 suites 1 1 n/a 0 0 00:36:03.599 tests 23 23 23 0 0 00:36:03.599 asserts 152 152 152 0 n/a 00:36:03.599 00:36:03.599 Elapsed time = 1.224 seconds 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@516 -- # nvmfcleanup 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:03.599 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:03.599 rmmod nvme_tcp 00:36:03.599 rmmod nvme_fabrics 00:36:03.860 rmmod nvme_keyring 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@517 -- # '[' -n 4162065 ']' 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@518 -- # killprocess 4162065 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@954 -- # '[' -z 4162065 ']' 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@958 -- # kill -0 4162065 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@959 -- # uname 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4162065 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@960 -- # process_name=reactor_3 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@964 -- # '[' reactor_3 = sudo ']' 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4162065' 00:36:03.860 killing process with pid 4162065 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@973 -- # kill 4162065 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@978 -- # wait 4162065 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-save 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:36:03.860 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-restore 00:36:04.121 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:04.121 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:04.121 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:04.121 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:04.121 19:32:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:06.032 19:32:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:06.032 00:36:06.032 real 0m12.852s 00:36:06.032 user 0m9.109s 00:36:06.032 sys 0m7.099s 00:36:06.032 19:32:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:06.032 19:32:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:36:06.032 ************************************ 00:36:06.032 END TEST nvmf_bdevio 00:36:06.032 ************************************ 00:36:06.032 19:32:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:36:06.032 00:36:06.032 real 5m10.003s 00:36:06.032 user 10m11.999s 00:36:06.032 sys 2m12.269s 00:36:06.032 19:32:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:06.032 19:32:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:36:06.032 ************************************ 00:36:06.032 END TEST nvmf_target_core_interrupt_mode 00:36:06.032 ************************************ 00:36:06.032 19:32:18 nvmf_tcp -- nvmf/nvmf.sh@21 -- # run_test nvmf_interrupt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:36:06.032 19:32:18 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:36:06.032 19:32:18 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:06.032 19:32:18 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:06.293 ************************************ 00:36:06.293 START TEST nvmf_interrupt 00:36:06.293 ************************************ 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:36:06.293 * Looking for test storage... 00:36:06.293 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1693 -- # lcov --version 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # IFS=.-: 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # read -ra ver1 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # IFS=.-: 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # read -ra ver2 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@338 -- # local 'op=<' 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@340 -- # ver1_l=2 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@341 -- # ver2_l=1 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@344 -- # case "$op" in 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@345 -- # : 1 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # decimal 1 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=1 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 1 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # ver1[v]=1 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # decimal 2 00:36:06.293 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=2 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 2 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # ver2[v]=2 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # return 0 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:36:06.294 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:06.294 --rc genhtml_branch_coverage=1 00:36:06.294 --rc genhtml_function_coverage=1 00:36:06.294 --rc genhtml_legend=1 00:36:06.294 --rc geninfo_all_blocks=1 00:36:06.294 --rc geninfo_unexecuted_blocks=1 00:36:06.294 00:36:06.294 ' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:36:06.294 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:06.294 --rc genhtml_branch_coverage=1 00:36:06.294 --rc genhtml_function_coverage=1 00:36:06.294 --rc genhtml_legend=1 00:36:06.294 --rc geninfo_all_blocks=1 00:36:06.294 --rc geninfo_unexecuted_blocks=1 00:36:06.294 00:36:06.294 ' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:36:06.294 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:06.294 --rc genhtml_branch_coverage=1 00:36:06.294 --rc genhtml_function_coverage=1 00:36:06.294 --rc genhtml_legend=1 00:36:06.294 --rc geninfo_all_blocks=1 00:36:06.294 --rc geninfo_unexecuted_blocks=1 00:36:06.294 00:36:06.294 ' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:36:06.294 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:06.294 --rc genhtml_branch_coverage=1 00:36:06.294 --rc genhtml_function_coverage=1 00:36:06.294 --rc genhtml_legend=1 00:36:06.294 --rc geninfo_all_blocks=1 00:36:06.294 --rc geninfo_unexecuted_blocks=1 00:36:06.294 00:36:06.294 ' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # uname -s 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@15 -- # shopt -s extglob 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@5 -- # export PATH 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@51 -- # : 0 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/interrupt/common.sh 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@12 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@14 -- # nvmftestinit 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@476 -- # prepare_net_devs 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@438 -- # local -g is_hw=no 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # remove_spdk_ns 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@309 -- # xtrace_disable 00:36:06.294 19:32:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:14.430 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:14.430 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # pci_devs=() 00:36:14.430 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:14.430 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:14.430 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:14.430 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:14.430 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # net_devs=() 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # e810=() 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # local -ga e810 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # x722=() 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # local -ga x722 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # mlx=() 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # local -ga mlx 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:36:14.431 Found 0000:31:00.0 (0x8086 - 0x159b) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:36:14.431 Found 0000:31:00.1 (0x8086 - 0x159b) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:36:14.431 Found net devices under 0000:31:00.0: cvl_0_0 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:36:14.431 Found net devices under 0000:31:00.1: cvl_0_1 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # is_hw=yes 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:14.431 19:32:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:14.692 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:14.692 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.603 ms 00:36:14.692 00:36:14.692 --- 10.0.0.2 ping statistics --- 00:36:14.692 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:14.692 rtt min/avg/max/mdev = 0.603/0.603/0.603/0.000 ms 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:14.692 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:14.692 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.324 ms 00:36:14.692 00:36:14.692 --- 10.0.0.1 ping statistics --- 00:36:14.692 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:14.692 rtt min/avg/max/mdev = 0.324/0.324/0.324/0.000 ms 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@450 -- # return 0 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@15 -- # nvmfappstart -m 0x3 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@509 -- # nvmfpid=4167301 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@510 -- # waitforlisten 4167301 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@835 -- # '[' -z 4167301 ']' 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:14.692 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:14.692 19:32:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:14.692 [2024-11-26 19:32:27.218946] thread.c:2977:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:36:14.692 [2024-11-26 19:32:27.220113] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:36:14.692 [2024-11-26 19:32:27.220168] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:14.692 [2024-11-26 19:32:27.310522] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:36:14.952 [2024-11-26 19:32:27.351007] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:14.952 [2024-11-26 19:32:27.351045] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:14.952 [2024-11-26 19:32:27.351053] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:14.952 [2024-11-26 19:32:27.351060] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:14.952 [2024-11-26 19:32:27.351066] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:14.952 [2024-11-26 19:32:27.352334] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:14.952 [2024-11-26 19:32:27.352336] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:14.952 [2024-11-26 19:32:27.409123] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:36:14.952 [2024-11-26 19:32:27.409636] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:36:14.952 [2024-11-26 19:32:27.409992] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@868 -- # return 0 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@16 -- # setup_bdev_aio 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # uname -s 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # [[ Linux != \F\r\e\e\B\S\D ]] 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@78 -- # dd if=/dev/zero of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile bs=2048 count=5000 00:36:15.524 5000+0 records in 00:36:15.524 5000+0 records out 00:36:15.524 10240000 bytes (10 MB, 9.8 MiB) copied, 0.0186725 s, 548 MB/s 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@79 -- # rpc_cmd bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile AIO0 2048 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:15.524 AIO0 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -q 256 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:15.524 [2024-11-26 19:32:28.109306] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 AIO0 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:15.524 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:15.785 [2024-11-26 19:32:28.149763] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 4167301 0 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 4167301 0 idle 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167301 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.26 reactor_0' 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167301 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.26 reactor_0 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 4167301 1 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 4167301 1 idle 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:15.785 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167355 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.00 reactor_1' 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167355 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.00 reactor_1 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@28 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@35 -- # perf_pid=4167490 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 256 -o 4096 -w randrw -M 30 -t 10 -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 4167301 0 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 4167301 0 busy 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:16.047 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167301 root 20 0 128.2g 43776 32256 R 33.3 0.0 0:00.31 reactor_0' 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167301 root 20 0 128.2g 43776 32256 R 33.3 0.0 0:00.31 reactor_0 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=33.3 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=33 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 4167301 1 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 4167301 1 busy 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167355 root 20 0 128.2g 43776 32256 R 99.9 0.0 0:00.21 reactor_1' 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167355 root 20 0 128.2g 43776 32256 R 99.9 0.0 0:00.21 reactor_1 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=99.9 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=99 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:16.308 19:32:28 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@42 -- # wait 4167490 00:36:26.315 Initializing NVMe Controllers 00:36:26.315 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:36:26.315 Controller IO queue size 256, less than required. 00:36:26.315 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:36:26.315 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:36:26.315 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:36:26.315 Initialization complete. Launching workers. 00:36:26.315 ======================================================== 00:36:26.315 Latency(us) 00:36:26.315 Device Information : IOPS MiB/s Average min max 00:36:26.315 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 16657.95 65.07 15377.78 2602.74 18921.54 00:36:26.315 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 19249.22 75.19 13301.05 7531.18 30719.26 00:36:26.315 ======================================================== 00:36:26.315 Total : 35907.17 140.26 14264.48 2602.74 30719.26 00:36:26.315 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 4167301 0 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 4167301 0 idle 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167301 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:20.27 reactor_0' 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167301 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:20.27 reactor_0 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 4167301 1 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 4167301 1 idle 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:26.315 19:32:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167355 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:10.00 reactor_1' 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167355 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:10.00 reactor_1 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:26.576 19:32:39 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@50 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:36:27.147 19:32:39 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@51 -- # waitforserial SPDKISFASTANDAWESOME 00:36:27.147 19:32:39 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1202 -- # local i=0 00:36:27.147 19:32:39 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:36:27.147 19:32:39 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:36:27.147 19:32:39 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1209 -- # sleep 2 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1212 -- # return 0 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 4167301 0 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 4167301 0 idle 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:29.063 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167301 root 20 0 128.2g 78336 32256 S 0.0 0.1 0:20.50 reactor_0' 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167301 root 20 0 128.2g 78336 32256 S 0.0 0.1 0:20.50 reactor_0 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 4167301 1 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 4167301 1 idle 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=4167301 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 4167301 -w 256 00:36:29.324 19:32:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='4167355 root 20 0 128.2g 78336 32256 S 0.0 0.1 0:10.13 reactor_1' 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 4167355 root 20 0 128.2g 78336 32256 S 0.0 0.1 0:10.13 reactor_1 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@55 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:36:29.585 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@56 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1223 -- # local i=0 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1235 -- # return 0 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@58 -- # trap - SIGINT SIGTERM EXIT 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@59 -- # nvmftestfini 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@516 -- # nvmfcleanup 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@121 -- # sync 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@124 -- # set +e 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:29.585 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:29.585 rmmod nvme_tcp 00:36:29.585 rmmod nvme_fabrics 00:36:29.846 rmmod nvme_keyring 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@128 -- # set -e 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@129 -- # return 0 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@517 -- # '[' -n 4167301 ']' 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@518 -- # killprocess 4167301 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@954 -- # '[' -z 4167301 ']' 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@958 -- # kill -0 4167301 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@959 -- # uname 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4167301 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4167301' 00:36:29.846 killing process with pid 4167301 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@973 -- # kill 4167301 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@978 -- # wait 4167301 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:36:29.846 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@297 -- # iptr 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@791 -- # iptables-save 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@791 -- # iptables-restore 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:36:30.106 19:32:42 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:32.019 19:32:44 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:32.020 00:36:32.020 real 0m25.888s 00:36:32.020 user 0m40.637s 00:36:32.020 sys 0m9.932s 00:36:32.020 19:32:44 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:32.020 19:32:44 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:36:32.020 ************************************ 00:36:32.020 END TEST nvmf_interrupt 00:36:32.020 ************************************ 00:36:32.020 00:36:32.020 real 31m14.233s 00:36:32.020 user 61m53.998s 00:36:32.020 sys 10m59.040s 00:36:32.020 19:32:44 nvmf_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:32.020 19:32:44 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:32.020 ************************************ 00:36:32.020 END TEST nvmf_tcp 00:36:32.020 ************************************ 00:36:32.020 19:32:44 -- spdk/autotest.sh@285 -- # [[ 0 -eq 0 ]] 00:36:32.020 19:32:44 -- spdk/autotest.sh@286 -- # run_test spdkcli_nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:36:32.020 19:32:44 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:36:32.020 19:32:44 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:32.020 19:32:44 -- common/autotest_common.sh@10 -- # set +x 00:36:32.281 ************************************ 00:36:32.281 START TEST spdkcli_nvmf_tcp 00:36:32.281 ************************************ 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:36:32.281 * Looking for test storage... 00:36:32.281 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:36:32.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:32.281 --rc genhtml_branch_coverage=1 00:36:32.281 --rc genhtml_function_coverage=1 00:36:32.281 --rc genhtml_legend=1 00:36:32.281 --rc geninfo_all_blocks=1 00:36:32.281 --rc geninfo_unexecuted_blocks=1 00:36:32.281 00:36:32.281 ' 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:36:32.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:32.281 --rc genhtml_branch_coverage=1 00:36:32.281 --rc genhtml_function_coverage=1 00:36:32.281 --rc genhtml_legend=1 00:36:32.281 --rc geninfo_all_blocks=1 00:36:32.281 --rc geninfo_unexecuted_blocks=1 00:36:32.281 00:36:32.281 ' 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:36:32.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:32.281 --rc genhtml_branch_coverage=1 00:36:32.281 --rc genhtml_function_coverage=1 00:36:32.281 --rc genhtml_legend=1 00:36:32.281 --rc geninfo_all_blocks=1 00:36:32.281 --rc geninfo_unexecuted_blocks=1 00:36:32.281 00:36:32.281 ' 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:36:32.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:32.281 --rc genhtml_branch_coverage=1 00:36:32.281 --rc genhtml_function_coverage=1 00:36:32.281 --rc genhtml_legend=1 00:36:32.281 --rc geninfo_all_blocks=1 00:36:32.281 --rc geninfo_unexecuted_blocks=1 00:36:32.281 00:36:32.281 ' 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # uname -s 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:32.281 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@15 -- # shopt -s extglob 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- paths/export.sh@5 -- # export PATH 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@51 -- # : 0 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:36:32.282 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@12 -- # MATCH_FILE=spdkcli_nvmf.test 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@13 -- # SPDKCLI_BRANCH=/nvmf 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@15 -- # trap cleanup EXIT 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@17 -- # timing_enter run_nvmf_tgt 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:32.282 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@18 -- # run_nvmf_tgt 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@33 -- # nvmf_tgt_pid=4170817 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@34 -- # waitforlisten 4170817 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@835 -- # '[' -z 4170817 ']' 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x3 -p 0 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:32.542 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:32.542 19:32:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:32.543 [2024-11-26 19:32:44.970250] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:36:32.543 [2024-11-26 19:32:44.970303] [ DPDK EAL parameters: nvmf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4170817 ] 00:36:32.543 [2024-11-26 19:32:45.048346] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:36:32.543 [2024-11-26 19:32:45.085733] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:32.543 [2024-11-26 19:32:45.085734] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@868 -- # return 0 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@19 -- # timing_exit run_nvmf_tgt 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@21 -- # NVMF_TARGET_IP=127.0.0.1 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@22 -- # [[ tcp == \r\d\m\a ]] 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@27 -- # timing_enter spdkcli_create_nvmf_config 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:33.482 19:32:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 32 512 Malloc1'\'' '\''Malloc1'\'' True 00:36:33.482 '\''/bdevs/malloc create 32 512 Malloc2'\'' '\''Malloc2'\'' True 00:36:33.482 '\''/bdevs/malloc create 32 512 Malloc3'\'' '\''Malloc3'\'' True 00:36:33.482 '\''/bdevs/malloc create 32 512 Malloc4'\'' '\''Malloc4'\'' True 00:36:33.482 '\''/bdevs/malloc create 32 512 Malloc5'\'' '\''Malloc5'\'' True 00:36:33.482 '\''/bdevs/malloc create 32 512 Malloc6'\'' '\''Malloc6'\'' True 00:36:33.482 '\''nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192'\'' '\'''\'' True 00:36:33.482 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1'\'' '\''Malloc3'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2'\'' '\''Malloc4'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:36:33.482 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2'\'' '\''Malloc2'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:36:33.482 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1'\'' '\''Malloc1'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True'\'' '\''Allow any host'\'' 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False'\'' '\''Allow any host'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4'\'' '\''127.0.0.1:4262'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5'\'' '\''Malloc5'\'' True 00:36:33.482 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6'\'' '\''Malloc6'\'' True 00:36:33.482 '\''/nvmf/referral create tcp 127.0.0.2 4030 IPv4'\'' 00:36:33.482 ' 00:36:36.026 [2024-11-26 19:32:48.211802] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:36.966 [2024-11-26 19:32:49.419744] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4260 *** 00:36:39.513 [2024-11-26 19:32:51.638387] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4261 *** 00:36:41.425 [2024-11-26 19:32:53.544004] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4262 *** 00:36:42.809 Executing command: ['/bdevs/malloc create 32 512 Malloc1', 'Malloc1', True] 00:36:42.809 Executing command: ['/bdevs/malloc create 32 512 Malloc2', 'Malloc2', True] 00:36:42.809 Executing command: ['/bdevs/malloc create 32 512 Malloc3', 'Malloc3', True] 00:36:42.809 Executing command: ['/bdevs/malloc create 32 512 Malloc4', 'Malloc4', True] 00:36:42.809 Executing command: ['/bdevs/malloc create 32 512 Malloc5', 'Malloc5', True] 00:36:42.809 Executing command: ['/bdevs/malloc create 32 512 Malloc6', 'Malloc6', True] 00:36:42.809 Executing command: ['nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192', '', True] 00:36:42.809 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode1', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1', 'Malloc3', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2', 'Malloc4', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:36:42.809 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2', 'Malloc2', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:36:42.809 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1', 'Malloc1', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1', 'nqn.2014-08.org.spdk:cnode1', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True', 'Allow any host', False] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False', 'Allow any host', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4', '127.0.0.1:4262', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5', 'Malloc5', True] 00:36:42.809 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6', 'Malloc6', True] 00:36:42.809 Executing command: ['/nvmf/referral create tcp 127.0.0.2 4030 IPv4', False] 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@66 -- # timing_exit spdkcli_create_nvmf_config 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@68 -- # timing_enter spdkcli_check_match 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@69 -- # check_match 00:36:42.809 19:32:55 spdkcli_nvmf_tcp -- spdkcli/common.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdkcli.py ll /nvmf 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- spdkcli/common.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/match/match /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test.match 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- spdkcli/common.sh@46 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@70 -- # timing_exit spdkcli_check_match 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@72 -- # timing_enter spdkcli_clear_nvmf_config 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:43.070 19:32:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1'\'' '\''Malloc3'\'' 00:36:43.070 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all'\'' '\''Malloc4'\'' 00:36:43.070 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:36:43.070 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' 00:36:43.070 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262'\'' '\''127.0.0.1:4262'\'' 00:36:43.070 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all'\'' '\''127.0.0.1:4261'\'' 00:36:43.070 '\''/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3'\'' '\''nqn.2014-08.org.spdk:cnode3'\'' 00:36:43.070 '\''/nvmf/subsystem delete_all'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:36:43.070 '\''/bdevs/malloc delete Malloc6'\'' '\''Malloc6'\'' 00:36:43.070 '\''/bdevs/malloc delete Malloc5'\'' '\''Malloc5'\'' 00:36:43.070 '\''/bdevs/malloc delete Malloc4'\'' '\''Malloc4'\'' 00:36:43.070 '\''/bdevs/malloc delete Malloc3'\'' '\''Malloc3'\'' 00:36:43.070 '\''/bdevs/malloc delete Malloc2'\'' '\''Malloc2'\'' 00:36:43.070 '\''/bdevs/malloc delete Malloc1'\'' '\''Malloc1'\'' 00:36:43.070 ' 00:36:48.358 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1', 'Malloc3', False] 00:36:48.358 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all', 'Malloc4', False] 00:36:48.358 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', False] 00:36:48.358 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all', 'nqn.2014-08.org.spdk:cnode1', False] 00:36:48.358 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262', '127.0.0.1:4262', False] 00:36:48.358 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all', '127.0.0.1:4261', False] 00:36:48.358 Executing command: ['/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3', 'nqn.2014-08.org.spdk:cnode3', False] 00:36:48.358 Executing command: ['/nvmf/subsystem delete_all', 'nqn.2014-08.org.spdk:cnode2', False] 00:36:48.358 Executing command: ['/bdevs/malloc delete Malloc6', 'Malloc6', False] 00:36:48.358 Executing command: ['/bdevs/malloc delete Malloc5', 'Malloc5', False] 00:36:48.358 Executing command: ['/bdevs/malloc delete Malloc4', 'Malloc4', False] 00:36:48.358 Executing command: ['/bdevs/malloc delete Malloc3', 'Malloc3', False] 00:36:48.358 Executing command: ['/bdevs/malloc delete Malloc2', 'Malloc2', False] 00:36:48.358 Executing command: ['/bdevs/malloc delete Malloc1', 'Malloc1', False] 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@88 -- # timing_exit spdkcli_clear_nvmf_config 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@90 -- # killprocess 4170817 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # '[' -z 4170817 ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@958 -- # kill -0 4170817 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@959 -- # uname 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4170817 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4170817' 00:36:48.358 killing process with pid 4170817 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@973 -- # kill 4170817 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@978 -- # wait 4170817 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@1 -- # cleanup 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@10 -- # '[' -n '' ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@13 -- # '[' -n 4170817 ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@14 -- # killprocess 4170817 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # '[' -z 4170817 ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@958 -- # kill -0 4170817 00:36:48.358 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (4170817) - No such process 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@981 -- # echo 'Process with pid 4170817 is not found' 00:36:48.358 Process with pid 4170817 is not found 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@22 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_nvmf.test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:36:48.358 00:36:48.358 real 0m16.225s 00:36:48.358 user 0m33.589s 00:36:48.358 sys 0m0.712s 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:48.358 19:33:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:36:48.358 ************************************ 00:36:48.358 END TEST spdkcli_nvmf_tcp 00:36:48.358 ************************************ 00:36:48.358 19:33:00 -- spdk/autotest.sh@287 -- # run_test nvmf_identify_passthru /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:36:48.358 19:33:00 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:36:48.358 19:33:00 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:48.358 19:33:00 -- common/autotest_common.sh@10 -- # set +x 00:36:48.358 ************************************ 00:36:48.358 START TEST nvmf_identify_passthru 00:36:48.358 ************************************ 00:36:48.358 19:33:00 nvmf_identify_passthru -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:36:48.620 * Looking for test storage... 00:36:48.620 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:36:48.620 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:36:48.620 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1693 -- # lcov --version 00:36:48.620 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:36:48.620 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@336 -- # IFS=.-: 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@336 -- # read -ra ver1 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@337 -- # IFS=.-: 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@337 -- # read -ra ver2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@338 -- # local 'op=<' 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@340 -- # ver1_l=2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@341 -- # ver2_l=1 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@344 -- # case "$op" in 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@345 -- # : 1 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@365 -- # decimal 1 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=1 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 1 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@365 -- # ver1[v]=1 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@366 -- # decimal 2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@366 -- # ver2[v]=2 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:48.620 19:33:01 nvmf_identify_passthru -- scripts/common.sh@368 -- # return 0 00:36:48.620 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:48.620 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:36:48.620 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.620 --rc genhtml_branch_coverage=1 00:36:48.620 --rc genhtml_function_coverage=1 00:36:48.620 --rc genhtml_legend=1 00:36:48.620 --rc geninfo_all_blocks=1 00:36:48.620 --rc geninfo_unexecuted_blocks=1 00:36:48.620 00:36:48.620 ' 00:36:48.620 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:36:48.620 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.620 --rc genhtml_branch_coverage=1 00:36:48.621 --rc genhtml_function_coverage=1 00:36:48.621 --rc genhtml_legend=1 00:36:48.621 --rc geninfo_all_blocks=1 00:36:48.621 --rc geninfo_unexecuted_blocks=1 00:36:48.621 00:36:48.621 ' 00:36:48.621 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:36:48.621 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.621 --rc genhtml_branch_coverage=1 00:36:48.621 --rc genhtml_function_coverage=1 00:36:48.621 --rc genhtml_legend=1 00:36:48.621 --rc geninfo_all_blocks=1 00:36:48.621 --rc geninfo_unexecuted_blocks=1 00:36:48.621 00:36:48.621 ' 00:36:48.621 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:36:48.621 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.621 --rc genhtml_branch_coverage=1 00:36:48.621 --rc genhtml_function_coverage=1 00:36:48.621 --rc genhtml_legend=1 00:36:48.621 --rc geninfo_all_blocks=1 00:36:48.621 --rc geninfo_unexecuted_blocks=1 00:36:48.621 00:36:48.621 ' 00:36:48.621 19:33:01 nvmf_identify_passthru -- target/identify_passthru.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@7 -- # uname -s 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@51 -- # : 0 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:36:48.621 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:48.621 19:33:01 nvmf_identify_passthru -- target/identify_passthru.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:48.621 19:33:01 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:36:48.621 19:33:01 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.621 19:33:01 nvmf_identify_passthru -- target/identify_passthru.sh@12 -- # nvmftestinit 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@476 -- # prepare_net_devs 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@438 -- # local -g is_hw=no 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@440 -- # remove_spdk_ns 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:48.621 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:36:48.621 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:36:48.621 19:33:01 nvmf_identify_passthru -- nvmf/common.sh@309 -- # xtrace_disable 00:36:48.621 19:33:01 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@315 -- # pci_devs=() 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@319 -- # net_devs=() 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@320 -- # e810=() 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@320 -- # local -ga e810 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@321 -- # x722=() 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@321 -- # local -ga x722 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@322 -- # mlx=() 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@322 -- # local -ga mlx 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:36:56.766 Found 0000:31:00.0 (0x8086 - 0x159b) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:36:56.766 Found 0000:31:00.1 (0x8086 - 0x159b) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:56.766 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:36:56.767 Found net devices under 0000:31:00.0: cvl_0_0 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:36:56.767 Found net devices under 0000:31:00.1: cvl_0_1 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@442 -- # is_hw=yes 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:56.767 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:57.028 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:57.028 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.709 ms 00:36:57.028 00:36:57.028 --- 10.0.0.2 ping statistics --- 00:36:57.028 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:57.028 rtt min/avg/max/mdev = 0.709/0.709/0.709/0.000 ms 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:57.029 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:57.029 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.333 ms 00:36:57.029 00:36:57.029 --- 10.0.0.1 ping statistics --- 00:36:57.029 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:57.029 rtt min/avg/max/mdev = 0.333/0.333/0.333/0.000 ms 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@450 -- # return 0 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:36:57.029 19:33:09 nvmf_identify_passthru -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:36:57.029 19:33:09 nvmf_identify_passthru -- target/identify_passthru.sh@14 -- # timing_enter nvme_identify 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:57.029 19:33:09 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # get_first_nvme_bdf 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1509 -- # bdfs=() 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1509 -- # local bdfs 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1510 -- # bdfs=($(get_nvme_bdfs)) 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1510 -- # get_nvme_bdfs 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1498 -- # bdfs=() 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1498 -- # local bdfs 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1499 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1500 -- # (( 1 == 0 )) 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:65:00.0 00:36:57.029 19:33:09 nvmf_identify_passthru -- common/autotest_common.sh@1512 -- # echo 0000:65:00.0 00:36:57.029 19:33:09 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # bdf=0000:65:00.0 00:36:57.029 19:33:09 nvmf_identify_passthru -- target/identify_passthru.sh@17 -- # '[' -z 0000:65:00.0 ']' 00:36:57.029 19:33:09 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:65:00.0' -i 0 00:36:57.029 19:33:09 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # grep 'Serial Number:' 00:36:57.029 19:33:09 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # awk '{print $3}' 00:36:57.599 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # nvme_serial_number=S64GNE0R605494 00:36:57.599 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:65:00.0' -i 0 00:36:57.599 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # grep 'Model Number:' 00:36:57.599 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # awk '{print $3}' 00:36:58.170 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # nvme_model_number=SAMSUNG 00:36:58.170 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@26 -- # timing_exit nvme_identify 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:58.170 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@28 -- # timing_enter start_nvmf_tgt 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:58.170 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@31 -- # nvmfpid=4178828 00:36:58.170 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:36:58.170 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@30 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:36:58.170 19:33:10 nvmf_identify_passthru -- target/identify_passthru.sh@35 -- # waitforlisten 4178828 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@835 -- # '[' -z 4178828 ']' 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:58.170 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:58.170 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:58.171 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:58.171 19:33:10 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:58.171 [2024-11-26 19:33:10.660223] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:36:58.171 [2024-11-26 19:33:10.660278] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:58.171 [2024-11-26 19:33:10.743331] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:36:58.171 [2024-11-26 19:33:10.780013] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:58.171 [2024-11-26 19:33:10.780048] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:58.171 [2024-11-26 19:33:10.780056] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:58.171 [2024-11-26 19:33:10.780064] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:58.171 [2024-11-26 19:33:10.780070] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:58.171 [2024-11-26 19:33:10.781703] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:58.171 [2024-11-26 19:33:10.781837] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:36:58.171 [2024-11-26 19:33:10.781995] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:58.171 [2024-11-26 19:33:10.781996] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@868 -- # return 0 00:36:59.116 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@36 -- # rpc_cmd -v nvmf_set_config --passthru-identify-ctrlr 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.116 INFO: Log level set to 20 00:36:59.116 INFO: Requests: 00:36:59.116 { 00:36:59.116 "jsonrpc": "2.0", 00:36:59.116 "method": "nvmf_set_config", 00:36:59.116 "id": 1, 00:36:59.116 "params": { 00:36:59.116 "admin_cmd_passthru": { 00:36:59.116 "identify_ctrlr": true 00:36:59.116 } 00:36:59.116 } 00:36:59.116 } 00:36:59.116 00:36:59.116 INFO: response: 00:36:59.116 { 00:36:59.116 "jsonrpc": "2.0", 00:36:59.116 "id": 1, 00:36:59.116 "result": true 00:36:59.116 } 00:36:59.116 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.116 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@37 -- # rpc_cmd -v framework_start_init 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.116 INFO: Setting log level to 20 00:36:59.116 INFO: Setting log level to 20 00:36:59.116 INFO: Log level set to 20 00:36:59.116 INFO: Log level set to 20 00:36:59.116 INFO: Requests: 00:36:59.116 { 00:36:59.116 "jsonrpc": "2.0", 00:36:59.116 "method": "framework_start_init", 00:36:59.116 "id": 1 00:36:59.116 } 00:36:59.116 00:36:59.116 INFO: Requests: 00:36:59.116 { 00:36:59.116 "jsonrpc": "2.0", 00:36:59.116 "method": "framework_start_init", 00:36:59.116 "id": 1 00:36:59.116 } 00:36:59.116 00:36:59.116 [2024-11-26 19:33:11.523525] nvmf_tgt.c: 462:nvmf_tgt_advance_state: *NOTICE*: Custom identify ctrlr handler enabled 00:36:59.116 INFO: response: 00:36:59.116 { 00:36:59.116 "jsonrpc": "2.0", 00:36:59.116 "id": 1, 00:36:59.116 "result": true 00:36:59.116 } 00:36:59.116 00:36:59.116 INFO: response: 00:36:59.116 { 00:36:59.116 "jsonrpc": "2.0", 00:36:59.116 "id": 1, 00:36:59.116 "result": true 00:36:59.116 } 00:36:59.116 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.116 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@38 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.116 INFO: Setting log level to 40 00:36:59.116 INFO: Setting log level to 40 00:36:59.116 INFO: Setting log level to 40 00:36:59.116 [2024-11-26 19:33:11.536866] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.116 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@39 -- # timing_exit start_nvmf_tgt 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.116 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@41 -- # rpc_cmd bdev_nvme_attach_controller -b Nvme0 -t PCIe -a 0000:65:00.0 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.116 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.432 Nvme0n1 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.432 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@42 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 1 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.432 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@43 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.432 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@44 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.432 [2024-11-26 19:33:11.936098] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.432 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@46 -- # rpc_cmd nvmf_get_subsystems 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.432 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:36:59.432 [ 00:36:59.432 { 00:36:59.432 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:36:59.432 "subtype": "Discovery", 00:36:59.432 "listen_addresses": [], 00:36:59.432 "allow_any_host": true, 00:36:59.432 "hosts": [] 00:36:59.432 }, 00:36:59.432 { 00:36:59.432 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:36:59.432 "subtype": "NVMe", 00:36:59.432 "listen_addresses": [ 00:36:59.432 { 00:36:59.432 "trtype": "TCP", 00:36:59.432 "adrfam": "IPv4", 00:36:59.432 "traddr": "10.0.0.2", 00:36:59.432 "trsvcid": "4420" 00:36:59.432 } 00:36:59.432 ], 00:36:59.432 "allow_any_host": true, 00:36:59.432 "hosts": [], 00:36:59.432 "serial_number": "SPDK00000000000001", 00:36:59.432 "model_number": "SPDK bdev Controller", 00:36:59.432 "max_namespaces": 1, 00:36:59.432 "min_cntlid": 1, 00:36:59.432 "max_cntlid": 65519, 00:36:59.432 "namespaces": [ 00:36:59.432 { 00:36:59.432 "nsid": 1, 00:36:59.432 "bdev_name": "Nvme0n1", 00:36:59.432 "name": "Nvme0n1", 00:36:59.433 "nguid": "3634473052605494002538450000002D", 00:36:59.433 "uuid": "36344730-5260-5494-0025-38450000002d" 00:36:59.433 } 00:36:59.433 ] 00:36:59.433 } 00:36:59.433 ] 00:36:59.433 19:33:11 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.433 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:36:59.433 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # grep 'Serial Number:' 00:36:59.433 19:33:11 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # awk '{print $3}' 00:36:59.756 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # nvmf_serial_number=S64GNE0R605494 00:36:59.756 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:36:59.756 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # grep 'Model Number:' 00:36:59.756 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # awk '{print $3}' 00:37:00.017 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # nvmf_model_number=SAMSUNG 00:37:00.017 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@63 -- # '[' S64GNE0R605494 '!=' S64GNE0R605494 ']' 00:37:00.017 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@68 -- # '[' SAMSUNG '!=' SAMSUNG ']' 00:37:00.017 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@73 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:00.017 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@75 -- # trap - SIGINT SIGTERM EXIT 00:37:00.017 19:33:12 nvmf_identify_passthru -- target/identify_passthru.sh@77 -- # nvmftestfini 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@516 -- # nvmfcleanup 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@121 -- # sync 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@124 -- # set +e 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@125 -- # for i in {1..20} 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:37:00.017 rmmod nvme_tcp 00:37:00.017 rmmod nvme_fabrics 00:37:00.017 rmmod nvme_keyring 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@128 -- # set -e 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@129 -- # return 0 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@517 -- # '[' -n 4178828 ']' 00:37:00.017 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@518 -- # killprocess 4178828 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@954 -- # '[' -z 4178828 ']' 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@958 -- # kill -0 4178828 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@959 -- # uname 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4178828 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4178828' 00:37:00.017 killing process with pid 4178828 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@973 -- # kill 4178828 00:37:00.017 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@978 -- # wait 4178828 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@297 -- # iptr 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@791 -- # iptables-save 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@791 -- # iptables-restore 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@302 -- # remove_spdk_ns 00:37:00.277 19:33:12 nvmf_identify_passthru -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:00.277 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:37:00.277 19:33:12 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:02.214 19:33:14 nvmf_identify_passthru -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:37:02.476 00:37:02.476 real 0m13.864s 00:37:02.476 user 0m10.510s 00:37:02.476 sys 0m7.088s 00:37:02.476 19:33:14 nvmf_identify_passthru -- common/autotest_common.sh@1130 -- # xtrace_disable 00:37:02.476 19:33:14 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:37:02.476 ************************************ 00:37:02.476 END TEST nvmf_identify_passthru 00:37:02.476 ************************************ 00:37:02.476 19:33:14 -- spdk/autotest.sh@289 -- # run_test nvmf_dif /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:37:02.476 19:33:14 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:37:02.476 19:33:14 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:37:02.476 19:33:14 -- common/autotest_common.sh@10 -- # set +x 00:37:02.476 ************************************ 00:37:02.476 START TEST nvmf_dif 00:37:02.476 ************************************ 00:37:02.476 19:33:14 nvmf_dif -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:37:02.476 * Looking for test storage... 00:37:02.476 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:37:02.476 19:33:15 nvmf_dif -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:37:02.476 19:33:15 nvmf_dif -- common/autotest_common.sh@1693 -- # lcov --version 00:37:02.476 19:33:15 nvmf_dif -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:37:02.476 19:33:15 nvmf_dif -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@333 -- # local ver1 ver1_l 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@334 -- # local ver2 ver2_l 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@336 -- # IFS=.-: 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@336 -- # read -ra ver1 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@337 -- # IFS=.-: 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@337 -- # read -ra ver2 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@338 -- # local 'op=<' 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@340 -- # ver1_l=2 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@341 -- # ver2_l=1 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@344 -- # case "$op" in 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@345 -- # : 1 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@364 -- # (( v = 0 )) 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:37:02.476 19:33:15 nvmf_dif -- scripts/common.sh@365 -- # decimal 1 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@353 -- # local d=1 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@355 -- # echo 1 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@365 -- # ver1[v]=1 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@366 -- # decimal 2 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@353 -- # local d=2 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@355 -- # echo 2 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@366 -- # ver2[v]=2 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@368 -- # return 0 00:37:02.737 19:33:15 nvmf_dif -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:37:02.737 19:33:15 nvmf_dif -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:37:02.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:02.737 --rc genhtml_branch_coverage=1 00:37:02.737 --rc genhtml_function_coverage=1 00:37:02.737 --rc genhtml_legend=1 00:37:02.737 --rc geninfo_all_blocks=1 00:37:02.737 --rc geninfo_unexecuted_blocks=1 00:37:02.737 00:37:02.737 ' 00:37:02.737 19:33:15 nvmf_dif -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:37:02.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:02.737 --rc genhtml_branch_coverage=1 00:37:02.737 --rc genhtml_function_coverage=1 00:37:02.737 --rc genhtml_legend=1 00:37:02.737 --rc geninfo_all_blocks=1 00:37:02.737 --rc geninfo_unexecuted_blocks=1 00:37:02.737 00:37:02.737 ' 00:37:02.737 19:33:15 nvmf_dif -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:37:02.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:02.737 --rc genhtml_branch_coverage=1 00:37:02.737 --rc genhtml_function_coverage=1 00:37:02.737 --rc genhtml_legend=1 00:37:02.737 --rc geninfo_all_blocks=1 00:37:02.737 --rc geninfo_unexecuted_blocks=1 00:37:02.737 00:37:02.737 ' 00:37:02.737 19:33:15 nvmf_dif -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:37:02.737 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:02.737 --rc genhtml_branch_coverage=1 00:37:02.737 --rc genhtml_function_coverage=1 00:37:02.737 --rc genhtml_legend=1 00:37:02.737 --rc geninfo_all_blocks=1 00:37:02.737 --rc geninfo_unexecuted_blocks=1 00:37:02.737 00:37:02.737 ' 00:37:02.737 19:33:15 nvmf_dif -- target/dif.sh@13 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@7 -- # uname -s 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:37:02.737 19:33:15 nvmf_dif -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@15 -- # shopt -s extglob 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:37:02.737 19:33:15 nvmf_dif -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:37:02.737 19:33:15 nvmf_dif -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:02.737 19:33:15 nvmf_dif -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:02.738 19:33:15 nvmf_dif -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:02.738 19:33:15 nvmf_dif -- paths/export.sh@5 -- # export PATH 00:37:02.738 19:33:15 nvmf_dif -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@51 -- # : 0 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:37:02.738 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@55 -- # have_pci_nics=0 00:37:02.738 19:33:15 nvmf_dif -- target/dif.sh@15 -- # NULL_META=16 00:37:02.738 19:33:15 nvmf_dif -- target/dif.sh@15 -- # NULL_BLOCK_SIZE=512 00:37:02.738 19:33:15 nvmf_dif -- target/dif.sh@15 -- # NULL_SIZE=64 00:37:02.738 19:33:15 nvmf_dif -- target/dif.sh@15 -- # NULL_DIF=1 00:37:02.738 19:33:15 nvmf_dif -- target/dif.sh@135 -- # nvmftestinit 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@476 -- # prepare_net_devs 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@438 -- # local -g is_hw=no 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@440 -- # remove_spdk_ns 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:02.738 19:33:15 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:37:02.738 19:33:15 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:37:02.738 19:33:15 nvmf_dif -- nvmf/common.sh@309 -- # xtrace_disable 00:37:02.738 19:33:15 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@315 -- # pci_devs=() 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@315 -- # local -a pci_devs 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@316 -- # pci_net_devs=() 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@317 -- # pci_drivers=() 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@317 -- # local -A pci_drivers 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@319 -- # net_devs=() 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@319 -- # local -ga net_devs 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@320 -- # e810=() 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@320 -- # local -ga e810 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@321 -- # x722=() 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@321 -- # local -ga x722 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@322 -- # mlx=() 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@322 -- # local -ga mlx 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:37:10.876 Found 0000:31:00.0 (0x8086 - 0x159b) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:37:10.876 Found 0000:31:00.1 (0x8086 - 0x159b) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@418 -- # [[ up == up ]] 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:37:10.876 Found net devices under 0000:31:00.0: cvl_0_0 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:37:10.876 19:33:22 nvmf_dif -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@418 -- # [[ up == up ]] 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:37:10.877 Found net devices under 0000:31:00.1: cvl_0_1 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@442 -- # is_hw=yes 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:37:10.877 19:33:22 nvmf_dif -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:37:10.877 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:37:10.877 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.631 ms 00:37:10.877 00:37:10.877 --- 10.0.0.2 ping statistics --- 00:37:10.877 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:10.877 rtt min/avg/max/mdev = 0.631/0.631/0.631/0.000 ms 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:37:10.877 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:37:10.877 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.268 ms 00:37:10.877 00:37:10.877 --- 10.0.0.1 ping statistics --- 00:37:10.877 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:10.877 rtt min/avg/max/mdev = 0.268/0.268/0.268/0.000 ms 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@450 -- # return 0 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@478 -- # '[' iso == iso ']' 00:37:10.877 19:33:23 nvmf_dif -- nvmf/common.sh@479 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:37:14.184 0000:80:01.6 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:80:01.7 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:80:01.4 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:80:01.5 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:80:01.2 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:80:01.3 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:80:01.0 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:80:01.1 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:00:01.6 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:65:00.0 (144d a80a): Already using the vfio-pci driver 00:37:14.184 0000:00:01.7 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:00:01.4 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:00:01.5 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:00:01.2 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:00:01.3 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:00:01.0 (8086 0b00): Already using the vfio-pci driver 00:37:14.184 0000:00:01.1 (8086 0b00): Already using the vfio-pci driver 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:37:14.754 19:33:27 nvmf_dif -- target/dif.sh@136 -- # NVMF_TRANSPORT_OPTS+=' --dif-insert-or-strip' 00:37:14.754 19:33:27 nvmf_dif -- target/dif.sh@137 -- # nvmfappstart 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@726 -- # xtrace_disable 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@509 -- # nvmfpid=4185591 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@510 -- # waitforlisten 4185591 00:37:14.754 19:33:27 nvmf_dif -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@835 -- # '[' -z 4185591 ']' 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:37:14.754 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:14.754 19:33:27 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:14.754 [2024-11-26 19:33:27.273206] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:37:14.754 [2024-11-26 19:33:27.273264] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:37:14.754 [2024-11-26 19:33:27.361520] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:15.015 [2024-11-26 19:33:27.401967] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:37:15.015 [2024-11-26 19:33:27.402001] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:37:15.015 [2024-11-26 19:33:27.402009] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:37:15.015 [2024-11-26 19:33:27.402016] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:37:15.015 [2024-11-26 19:33:27.402022] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:37:15.015 [2024-11-26 19:33:27.402639] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@868 -- # return 0 00:37:15.619 19:33:28 nvmf_dif -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@732 -- # xtrace_disable 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:15.619 19:33:28 nvmf_dif -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:37:15.619 19:33:28 nvmf_dif -- target/dif.sh@139 -- # create_transport 00:37:15.619 19:33:28 nvmf_dif -- target/dif.sh@50 -- # rpc_cmd nvmf_create_transport -t tcp -o --dif-insert-or-strip 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:15.619 [2024-11-26 19:33:28.104926] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.619 19:33:28 nvmf_dif -- target/dif.sh@141 -- # run_test fio_dif_1_default fio_dif_1 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:37:15.619 19:33:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:15.619 ************************************ 00:37:15.619 START TEST fio_dif_1_default 00:37:15.619 ************************************ 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1129 -- # fio_dif_1 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@86 -- # create_subsystems 0 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@28 -- # local sub 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@30 -- # for sub in "$@" 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@31 -- # create_subsystem 0 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@18 -- # local sub_id=0 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:37:15.619 bdev_null0 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:37:15.619 [2024-11-26 19:33:28.177244] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # fio /dev/fd/62 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # create_json_sub_conf 0 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@560 -- # config=() 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@560 -- # local subsystem config 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:15.619 { 00:37:15.619 "params": { 00:37:15.619 "name": "Nvme$subsystem", 00:37:15.619 "trtype": "$TEST_TRANSPORT", 00:37:15.619 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:15.619 "adrfam": "ipv4", 00:37:15.619 "trsvcid": "$NVMF_PORT", 00:37:15.619 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:15.619 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:15.619 "hdgst": ${hdgst:-false}, 00:37:15.619 "ddgst": ${ddgst:-false} 00:37:15.619 }, 00:37:15.619 "method": "bdev_nvme_attach_controller" 00:37:15.619 } 00:37:15.619 EOF 00:37:15.619 )") 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # gen_fio_conf 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@54 -- # local file 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1343 -- # local sanitizers 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@56 -- # cat 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # shift 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1347 -- # local asan_lib= 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@582 -- # cat 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file = 1 )) 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # grep libasan 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file <= files )) 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@584 -- # jq . 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@585 -- # IFS=, 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:37:15.619 "params": { 00:37:15.619 "name": "Nvme0", 00:37:15.619 "trtype": "tcp", 00:37:15.619 "traddr": "10.0.0.2", 00:37:15.619 "adrfam": "ipv4", 00:37:15.619 "trsvcid": "4420", 00:37:15.619 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:15.619 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:15.619 "hdgst": false, 00:37:15.619 "ddgst": false 00:37:15.619 }, 00:37:15.619 "method": "bdev_nvme_attach_controller" 00:37:15.619 }' 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:37:15.619 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:15.901 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:15.901 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:15.901 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:37:15.901 19:33:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:16.164 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:37:16.164 fio-3.35 00:37:16.164 Starting 1 thread 00:37:28.407 00:37:28.407 filename0: (groupid=0, jobs=1): err= 0: pid=4186126: Tue Nov 26 19:33:39 2024 00:37:28.407 read: IOPS=96, BW=388KiB/s (397kB/s)(3888KiB/10030msec) 00:37:28.407 slat (nsec): min=5496, max=40760, avg=6427.89, stdev=1868.78 00:37:28.407 clat (usec): min=942, max=45313, avg=41255.44, stdev=2650.51 00:37:28.407 lat (usec): min=950, max=45353, avg=41261.86, stdev=2650.48 00:37:28.407 clat percentiles (usec): 00:37:28.407 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:37:28.407 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41681], 00:37:28.407 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:37:28.407 | 99.00th=[42730], 99.50th=[42730], 99.90th=[45351], 99.95th=[45351], 00:37:28.407 | 99.99th=[45351] 00:37:28.407 bw ( KiB/s): min= 384, max= 416, per=99.84%, avg=387.20, stdev= 9.85, samples=20 00:37:28.407 iops : min= 96, max= 104, avg=96.80, stdev= 2.46, samples=20 00:37:28.407 lat (usec) : 1000=0.41% 00:37:28.407 lat (msec) : 50=99.59% 00:37:28.407 cpu : usr=93.25%, sys=6.53%, ctx=13, majf=0, minf=249 00:37:28.407 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:28.407 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:28.407 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:28.407 issued rwts: total=972,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:28.407 latency : target=0, window=0, percentile=100.00%, depth=4 00:37:28.407 00:37:28.407 Run status group 0 (all jobs): 00:37:28.407 READ: bw=388KiB/s (397kB/s), 388KiB/s-388KiB/s (397kB/s-397kB/s), io=3888KiB (3981kB), run=10030-10030msec 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@88 -- # destroy_subsystems 0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@43 -- # local sub 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@45 -- # for sub in "$@" 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@46 -- # destroy_subsystem 0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@36 -- # local sub_id=0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 00:37:28.407 real 0m11.166s 00:37:28.407 user 0m26.046s 00:37:28.407 sys 0m0.977s 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1130 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 ************************************ 00:37:28.407 END TEST fio_dif_1_default 00:37:28.407 ************************************ 00:37:28.407 19:33:39 nvmf_dif -- target/dif.sh@142 -- # run_test fio_dif_1_multi_subsystems fio_dif_1_multi_subsystems 00:37:28.407 19:33:39 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:37:28.407 19:33:39 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 ************************************ 00:37:28.407 START TEST fio_dif_1_multi_subsystems 00:37:28.407 ************************************ 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1129 -- # fio_dif_1_multi_subsystems 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@92 -- # local files=1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@94 -- # create_subsystems 0 1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@28 -- # local sub 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 bdev_null0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 [2024-11-26 19:33:39.427532] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 bdev_null1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # fio /dev/fd/62 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # create_json_sub_conf 0 1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # config=() 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # local subsystem config 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:28.407 { 00:37:28.407 "params": { 00:37:28.407 "name": "Nvme$subsystem", 00:37:28.407 "trtype": "$TEST_TRANSPORT", 00:37:28.407 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:28.407 "adrfam": "ipv4", 00:37:28.407 "trsvcid": "$NVMF_PORT", 00:37:28.407 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:28.407 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:28.407 "hdgst": ${hdgst:-false}, 00:37:28.407 "ddgst": ${ddgst:-false} 00:37:28.407 }, 00:37:28.407 "method": "bdev_nvme_attach_controller" 00:37:28.407 } 00:37:28.407 EOF 00:37:28.407 )") 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # gen_fio_conf 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@54 -- # local file 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1343 -- # local sanitizers 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@56 -- # cat 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # shift 00:37:28.407 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1347 -- # local asan_lib= 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # cat 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # grep libasan 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file = 1 )) 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@73 -- # cat 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:28.408 { 00:37:28.408 "params": { 00:37:28.408 "name": "Nvme$subsystem", 00:37:28.408 "trtype": "$TEST_TRANSPORT", 00:37:28.408 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:28.408 "adrfam": "ipv4", 00:37:28.408 "trsvcid": "$NVMF_PORT", 00:37:28.408 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:28.408 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:28.408 "hdgst": ${hdgst:-false}, 00:37:28.408 "ddgst": ${ddgst:-false} 00:37:28.408 }, 00:37:28.408 "method": "bdev_nvme_attach_controller" 00:37:28.408 } 00:37:28.408 EOF 00:37:28.408 )") 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file++ )) 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # cat 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@584 -- # jq . 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@585 -- # IFS=, 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:37:28.408 "params": { 00:37:28.408 "name": "Nvme0", 00:37:28.408 "trtype": "tcp", 00:37:28.408 "traddr": "10.0.0.2", 00:37:28.408 "adrfam": "ipv4", 00:37:28.408 "trsvcid": "4420", 00:37:28.408 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:28.408 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:28.408 "hdgst": false, 00:37:28.408 "ddgst": false 00:37:28.408 }, 00:37:28.408 "method": "bdev_nvme_attach_controller" 00:37:28.408 },{ 00:37:28.408 "params": { 00:37:28.408 "name": "Nvme1", 00:37:28.408 "trtype": "tcp", 00:37:28.408 "traddr": "10.0.0.2", 00:37:28.408 "adrfam": "ipv4", 00:37:28.408 "trsvcid": "4420", 00:37:28.408 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:37:28.408 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:37:28.408 "hdgst": false, 00:37:28.408 "ddgst": false 00:37:28.408 }, 00:37:28.408 "method": "bdev_nvme_attach_controller" 00:37:28.408 }' 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:37:28.408 19:33:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:28.408 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:37:28.408 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:37:28.408 fio-3.35 00:37:28.408 Starting 2 threads 00:37:38.411 00:37:38.411 filename0: (groupid=0, jobs=1): err= 0: pid=4188550: Tue Nov 26 19:33:50 2024 00:37:38.411 read: IOPS=96, BW=386KiB/s (396kB/s)(3872KiB/10023msec) 00:37:38.411 slat (nsec): min=5481, max=53163, avg=6989.64, stdev=3125.50 00:37:38.411 clat (usec): min=40759, max=43067, avg=41395.62, stdev=556.67 00:37:38.411 lat (usec): min=40764, max=43078, avg=41402.61, stdev=557.34 00:37:38.411 clat percentiles (usec): 00:37:38.411 | 1.00th=[40633], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:37:38.411 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:37:38.411 | 70.00th=[41681], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:37:38.411 | 99.00th=[43254], 99.50th=[43254], 99.90th=[43254], 99.95th=[43254], 00:37:38.411 | 99.99th=[43254] 00:37:38.411 bw ( KiB/s): min= 352, max= 416, per=49.83%, avg=385.60, stdev=12.61, samples=20 00:37:38.411 iops : min= 88, max= 104, avg=96.40, stdev= 3.15, samples=20 00:37:38.411 lat (msec) : 50=100.00% 00:37:38.411 cpu : usr=95.29%, sys=4.47%, ctx=16, majf=0, minf=178 00:37:38.411 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:38.411 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:38.411 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:38.411 issued rwts: total=968,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:38.411 latency : target=0, window=0, percentile=100.00%, depth=4 00:37:38.411 filename1: (groupid=0, jobs=1): err= 0: pid=4188551: Tue Nov 26 19:33:50 2024 00:37:38.412 read: IOPS=96, BW=386KiB/s (396kB/s)(3872KiB/10023msec) 00:37:38.412 slat (nsec): min=5479, max=49948, avg=6972.09, stdev=3003.85 00:37:38.412 clat (usec): min=865, max=43279, avg=41395.19, stdev=2688.42 00:37:38.412 lat (usec): min=873, max=43311, avg=41402.16, stdev=2687.91 00:37:38.412 clat percentiles (usec): 00:37:38.412 | 1.00th=[40633], 5.00th=[40633], 10.00th=[41157], 20.00th=[41157], 00:37:38.412 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41681], 00:37:38.412 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42730], 00:37:38.412 | 99.00th=[43254], 99.50th=[43254], 99.90th=[43254], 99.95th=[43254], 00:37:38.412 | 99.99th=[43254] 00:37:38.412 bw ( KiB/s): min= 352, max= 416, per=49.83%, avg=385.60, stdev=12.61, samples=20 00:37:38.412 iops : min= 88, max= 104, avg=96.40, stdev= 3.15, samples=20 00:37:38.412 lat (usec) : 1000=0.41% 00:37:38.412 lat (msec) : 50=99.59% 00:37:38.412 cpu : usr=95.33%, sys=4.44%, ctx=15, majf=0, minf=56 00:37:38.412 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:38.412 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:38.412 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:38.412 issued rwts: total=968,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:38.412 latency : target=0, window=0, percentile=100.00%, depth=4 00:37:38.412 00:37:38.412 Run status group 0 (all jobs): 00:37:38.412 READ: bw=773KiB/s (791kB/s), 386KiB/s-386KiB/s (396kB/s-396kB/s), io=7744KiB (7930kB), run=10023-10023msec 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@96 -- # destroy_subsystems 0 1 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@43 -- # local sub 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 1 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=1 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 00:37:38.412 real 0m11.506s 00:37:38.412 user 0m32.274s 00:37:38.412 sys 0m1.253s 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1130 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 ************************************ 00:37:38.412 END TEST fio_dif_1_multi_subsystems 00:37:38.412 ************************************ 00:37:38.412 19:33:50 nvmf_dif -- target/dif.sh@143 -- # run_test fio_dif_rand_params fio_dif_rand_params 00:37:38.412 19:33:50 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:37:38.412 19:33:50 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 ************************************ 00:37:38.412 START TEST fio_dif_rand_params 00:37:38.412 ************************************ 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1129 -- # fio_dif_rand_params 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@100 -- # local NULL_DIF 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@101 -- # local bs numjobs runtime iodepth files 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # NULL_DIF=3 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # bs=128k 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # numjobs=3 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # iodepth=3 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # runtime=5 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@105 -- # create_subsystems 0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 bdev_null0 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:38.412 [2024-11-26 19:33:51.013716] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # fio /dev/fd/62 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # create_json_sub_conf 0 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:38.412 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # config=() 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # local subsystem config 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:38.413 { 00:37:38.413 "params": { 00:37:38.413 "name": "Nvme$subsystem", 00:37:38.413 "trtype": "$TEST_TRANSPORT", 00:37:38.413 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:38.413 "adrfam": "ipv4", 00:37:38.413 "trsvcid": "$NVMF_PORT", 00:37:38.413 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:38.413 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:38.413 "hdgst": ${hdgst:-false}, 00:37:38.413 "ddgst": ${ddgst:-false} 00:37:38.413 }, 00:37:38.413 "method": "bdev_nvme_attach_controller" 00:37:38.413 } 00:37:38.413 EOF 00:37:38.413 )") 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local sanitizers 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # shift 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1347 -- # local asan_lib= 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libasan 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # jq . 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@585 -- # IFS=, 00:37:38.413 19:33:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:37:38.413 "params": { 00:37:38.413 "name": "Nvme0", 00:37:38.413 "trtype": "tcp", 00:37:38.413 "traddr": "10.0.0.2", 00:37:38.413 "adrfam": "ipv4", 00:37:38.413 "trsvcid": "4420", 00:37:38.413 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:38.413 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:38.413 "hdgst": false, 00:37:38.413 "ddgst": false 00:37:38.413 }, 00:37:38.413 "method": "bdev_nvme_attach_controller" 00:37:38.413 }' 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:37:38.673 19:33:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:38.933 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:37:38.933 ... 00:37:38.933 fio-3.35 00:37:38.933 Starting 3 threads 00:37:45.517 00:37:45.517 filename0: (groupid=0, jobs=1): err= 0: pid=4190842: Tue Nov 26 19:33:57 2024 00:37:45.517 read: IOPS=196, BW=24.5MiB/s (25.7MB/s)(124MiB/5047msec) 00:37:45.517 slat (nsec): min=2882, max=13402, avg=6251.04, stdev=496.62 00:37:45.517 clat (usec): min=6425, max=55718, avg=15227.59, stdev=11034.05 00:37:45.517 lat (usec): min=6431, max=55724, avg=15233.84, stdev=11034.04 00:37:45.517 clat percentiles (usec): 00:37:45.517 | 1.00th=[ 8029], 5.00th=[ 9241], 10.00th=[ 9896], 20.00th=[10683], 00:37:45.517 | 30.00th=[11076], 40.00th=[11600], 50.00th=[12125], 60.00th=[12518], 00:37:45.517 | 70.00th=[13042], 80.00th=[14091], 90.00th=[16057], 95.00th=[51119], 00:37:45.517 | 99.00th=[53216], 99.50th=[55313], 99.90th=[55837], 99.95th=[55837], 00:37:45.518 | 99.99th=[55837] 00:37:45.518 bw ( KiB/s): min=19712, max=30976, per=30.08%, avg=25292.80, stdev=3730.93, samples=10 00:37:45.518 iops : min= 154, max= 242, avg=197.60, stdev=29.15, samples=10 00:37:45.518 lat (msec) : 10=11.30%, 20=80.32%, 50=2.22%, 100=6.16% 00:37:45.518 cpu : usr=94.71%, sys=5.07%, ctx=10, majf=0, minf=56 00:37:45.518 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:45.518 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:45.518 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:45.518 issued rwts: total=991,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:45.518 latency : target=0, window=0, percentile=100.00%, depth=3 00:37:45.518 filename0: (groupid=0, jobs=1): err= 0: pid=4190843: Tue Nov 26 19:33:57 2024 00:37:45.518 read: IOPS=230, BW=28.9MiB/s (30.3MB/s)(146MiB/5046msec) 00:37:45.518 slat (usec): min=5, max=157, avg= 8.00, stdev= 4.72 00:37:45.518 clat (usec): min=6234, max=54816, avg=12945.16, stdev=6642.95 00:37:45.518 lat (usec): min=6243, max=54825, avg=12953.15, stdev=6642.99 00:37:45.518 clat percentiles (usec): 00:37:45.518 | 1.00th=[ 7504], 5.00th=[ 8586], 10.00th=[ 9241], 20.00th=[10028], 00:37:45.518 | 30.00th=[10683], 40.00th=[11469], 50.00th=[12125], 60.00th=[12780], 00:37:45.518 | 70.00th=[13304], 80.00th=[13829], 90.00th=[14746], 95.00th=[15664], 00:37:45.518 | 99.00th=[51643], 99.50th=[52167], 99.90th=[54264], 99.95th=[54789], 00:37:45.518 | 99.99th=[54789] 00:37:45.518 bw ( KiB/s): min=23552, max=33024, per=35.41%, avg=29772.80, stdev=3007.44, samples=10 00:37:45.518 iops : min= 184, max= 258, avg=232.60, stdev=23.50, samples=10 00:37:45.518 lat (msec) : 10=19.31%, 20=77.94%, 50=1.12%, 100=1.63% 00:37:45.518 cpu : usr=95.44%, sys=4.30%, ctx=10, majf=0, minf=143 00:37:45.518 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:45.518 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:45.518 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:45.518 issued rwts: total=1165,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:45.518 latency : target=0, window=0, percentile=100.00%, depth=3 00:37:45.518 filename0: (groupid=0, jobs=1): err= 0: pid=4190844: Tue Nov 26 19:33:57 2024 00:37:45.518 read: IOPS=229, BW=28.7MiB/s (30.1MB/s)(145MiB/5045msec) 00:37:45.518 slat (nsec): min=5768, max=44003, avg=8573.93, stdev=1588.50 00:37:45.518 clat (usec): min=6232, max=90270, avg=13010.76, stdev=6669.18 00:37:45.518 lat (usec): min=6240, max=90279, avg=13019.33, stdev=6669.23 00:37:45.518 clat percentiles (usec): 00:37:45.518 | 1.00th=[ 6521], 5.00th=[ 7701], 10.00th=[ 8586], 20.00th=[10028], 00:37:45.518 | 30.00th=[10814], 40.00th=[11731], 50.00th=[12256], 60.00th=[12911], 00:37:45.518 | 70.00th=[13566], 80.00th=[14353], 90.00th=[15401], 95.00th=[16450], 00:37:45.518 | 99.00th=[51119], 99.50th=[53216], 99.90th=[56886], 99.95th=[90702], 00:37:45.518 | 99.99th=[90702] 00:37:45.518 bw ( KiB/s): min=24320, max=33280, per=35.24%, avg=29624.90, stdev=3126.62, samples=10 00:37:45.518 iops : min= 190, max= 260, avg=231.40, stdev=24.44, samples=10 00:37:45.518 lat (msec) : 10=20.36%, 20=77.22%, 50=1.21%, 100=1.21% 00:37:45.518 cpu : usr=94.96%, sys=4.80%, ctx=7, majf=0, minf=78 00:37:45.518 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:45.518 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:45.518 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:45.518 issued rwts: total=1159,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:45.518 latency : target=0, window=0, percentile=100.00%, depth=3 00:37:45.518 00:37:45.518 Run status group 0 (all jobs): 00:37:45.518 READ: bw=82.1MiB/s (86.1MB/s), 24.5MiB/s-28.9MiB/s (25.7MB/s-30.3MB/s), io=414MiB (435MB), run=5045-5047msec 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@107 -- # destroy_subsystems 0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # NULL_DIF=2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # bs=4k 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # numjobs=8 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # iodepth=16 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # runtime= 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # files=2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@111 -- # create_subsystems 0 1 2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 bdev_null0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 [2024-11-26 19:33:57.283812] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 bdev_null1 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null2 64 512 --md-size 16 --dif-type 2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.518 bdev_null2 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 --serial-number 53313233-2 --allow-any-host 00:37:45.518 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 bdev_null2 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # fio /dev/fd/62 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # create_json_sub_conf 0 1 2 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 2 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # config=() 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # local subsystem config 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:45.519 { 00:37:45.519 "params": { 00:37:45.519 "name": "Nvme$subsystem", 00:37:45.519 "trtype": "$TEST_TRANSPORT", 00:37:45.519 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:45.519 "adrfam": "ipv4", 00:37:45.519 "trsvcid": "$NVMF_PORT", 00:37:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:45.519 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:45.519 "hdgst": ${hdgst:-false}, 00:37:45.519 "ddgst": ${ddgst:-false} 00:37:45.519 }, 00:37:45.519 "method": "bdev_nvme_attach_controller" 00:37:45.519 } 00:37:45.519 EOF 00:37:45.519 )") 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local sanitizers 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # shift 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1347 -- # local asan_lib= 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libasan 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:45.519 { 00:37:45.519 "params": { 00:37:45.519 "name": "Nvme$subsystem", 00:37:45.519 "trtype": "$TEST_TRANSPORT", 00:37:45.519 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:45.519 "adrfam": "ipv4", 00:37:45.519 "trsvcid": "$NVMF_PORT", 00:37:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:45.519 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:45.519 "hdgst": ${hdgst:-false}, 00:37:45.519 "ddgst": ${ddgst:-false} 00:37:45.519 }, 00:37:45.519 "method": "bdev_nvme_attach_controller" 00:37:45.519 } 00:37:45.519 EOF 00:37:45.519 )") 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:45.519 { 00:37:45.519 "params": { 00:37:45.519 "name": "Nvme$subsystem", 00:37:45.519 "trtype": "$TEST_TRANSPORT", 00:37:45.519 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:45.519 "adrfam": "ipv4", 00:37:45.519 "trsvcid": "$NVMF_PORT", 00:37:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:45.519 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:45.519 "hdgst": ${hdgst:-false}, 00:37:45.519 "ddgst": ${ddgst:-false} 00:37:45.519 }, 00:37:45.519 "method": "bdev_nvme_attach_controller" 00:37:45.519 } 00:37:45.519 EOF 00:37:45.519 )") 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # jq . 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@585 -- # IFS=, 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:37:45.519 "params": { 00:37:45.519 "name": "Nvme0", 00:37:45.519 "trtype": "tcp", 00:37:45.519 "traddr": "10.0.0.2", 00:37:45.519 "adrfam": "ipv4", 00:37:45.519 "trsvcid": "4420", 00:37:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:45.519 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:45.519 "hdgst": false, 00:37:45.519 "ddgst": false 00:37:45.519 }, 00:37:45.519 "method": "bdev_nvme_attach_controller" 00:37:45.519 },{ 00:37:45.519 "params": { 00:37:45.519 "name": "Nvme1", 00:37:45.519 "trtype": "tcp", 00:37:45.519 "traddr": "10.0.0.2", 00:37:45.519 "adrfam": "ipv4", 00:37:45.519 "trsvcid": "4420", 00:37:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:37:45.519 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:37:45.519 "hdgst": false, 00:37:45.519 "ddgst": false 00:37:45.519 }, 00:37:45.519 "method": "bdev_nvme_attach_controller" 00:37:45.519 },{ 00:37:45.519 "params": { 00:37:45.519 "name": "Nvme2", 00:37:45.519 "trtype": "tcp", 00:37:45.519 "traddr": "10.0.0.2", 00:37:45.519 "adrfam": "ipv4", 00:37:45.519 "trsvcid": "4420", 00:37:45.519 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:37:45.519 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:37:45.519 "hdgst": false, 00:37:45.519 "ddgst": false 00:37:45.519 }, 00:37:45.519 "method": "bdev_nvme_attach_controller" 00:37:45.519 }' 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:37:45.519 19:33:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:45.519 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:37:45.519 ... 00:37:45.519 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:37:45.519 ... 00:37:45.519 filename2: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:37:45.519 ... 00:37:45.519 fio-3.35 00:37:45.519 Starting 24 threads 00:37:57.758 00:37:57.758 filename0: (groupid=0, jobs=1): err= 0: pid=4192352: Tue Nov 26 19:34:09 2024 00:37:57.758 read: IOPS=485, BW=1942KiB/s (1989kB/s)(19.0MiB/10016msec) 00:37:57.758 slat (nsec): min=5854, max=64155, avg=18008.66, stdev=9865.07 00:37:57.758 clat (usec): min=7548, max=47257, avg=32782.69, stdev=2313.24 00:37:57.758 lat (usec): min=7564, max=47266, avg=32800.70, stdev=2313.53 00:37:57.758 clat percentiles (usec): 00:37:57.758 | 1.00th=[16909], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.758 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.758 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33424], 95.00th=[33817], 00:37:57.758 | 99.00th=[34341], 99.50th=[34866], 99.90th=[35390], 99.95th=[35390], 00:37:57.758 | 99.99th=[47449] 00:37:57.758 bw ( KiB/s): min= 1920, max= 2176, per=4.17%, avg=1939.20, stdev=62.64, samples=20 00:37:57.758 iops : min= 480, max= 544, avg=484.80, stdev=15.66, samples=20 00:37:57.759 lat (msec) : 10=0.06%, 20=1.30%, 50=98.64% 00:37:57.759 cpu : usr=98.70%, sys=0.93%, ctx=106, majf=0, minf=29 00:37:57.759 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=4864,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename0: (groupid=0, jobs=1): err= 0: pid=4192353: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=481, BW=1927KiB/s (1973kB/s)(18.8MiB/10011msec) 00:37:57.759 slat (nsec): min=5424, max=74564, avg=18000.73, stdev=11272.58 00:37:57.759 clat (usec): min=18795, max=56985, avg=33065.71, stdev=2266.98 00:37:57.759 lat (usec): min=18808, max=57000, avg=33083.71, stdev=2266.53 00:37:57.759 clat percentiles (usec): 00:37:57.759 | 1.00th=[25035], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.759 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.759 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:37:57.759 | 99.00th=[40109], 99.50th=[46400], 99.90th=[56886], 99.95th=[56886], 00:37:57.759 | 99.99th=[56886] 00:37:57.759 bw ( KiB/s): min= 1760, max= 2048, per=4.13%, avg=1922.53, stdev=51.78, samples=19 00:37:57.759 iops : min= 440, max= 512, avg=480.63, stdev=12.95, samples=19 00:37:57.759 lat (msec) : 20=0.41%, 50=99.25%, 100=0.33% 00:37:57.759 cpu : usr=98.76%, sys=0.87%, ctx=112, majf=0, minf=33 00:37:57.759 IO depths : 1=5.5%, 2=11.5%, 4=24.1%, 8=51.8%, 16=7.1%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=93.9%, 8=0.4%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=4822,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename0: (groupid=0, jobs=1): err= 0: pid=4192354: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=482, BW=1930KiB/s (1976kB/s)(18.9MiB/10017msec) 00:37:57.759 slat (nsec): min=5943, max=62510, avg=14682.79, stdev=8769.28 00:37:57.759 clat (usec): min=22153, max=41735, avg=33039.80, stdev=911.10 00:37:57.759 lat (usec): min=22174, max=41749, avg=33054.48, stdev=910.93 00:37:57.759 clat percentiles (usec): 00:37:57.759 | 1.00th=[31589], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.759 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.759 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.759 | 99.00th=[34866], 99.50th=[34866], 99.90th=[40109], 99.95th=[40633], 00:37:57.759 | 99.99th=[41681] 00:37:57.759 bw ( KiB/s): min= 1920, max= 2048, per=4.14%, avg=1926.74, stdev=29.37, samples=19 00:37:57.759 iops : min= 480, max= 512, avg=481.68, stdev= 7.34, samples=19 00:37:57.759 lat (msec) : 50=100.00% 00:37:57.759 cpu : usr=98.76%, sys=0.93%, ctx=77, majf=0, minf=39 00:37:57.759 IO depths : 1=6.0%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.5%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename0: (groupid=0, jobs=1): err= 0: pid=4192355: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=481, BW=1925KiB/s (1971kB/s)(18.8MiB/10006msec) 00:37:57.759 slat (nsec): min=5716, max=88725, avg=25389.67, stdev=15123.82 00:37:57.759 clat (usec): min=18738, max=60862, avg=33003.07, stdev=1914.59 00:37:57.759 lat (usec): min=18774, max=60877, avg=33028.46, stdev=1914.24 00:37:57.759 clat percentiles (usec): 00:37:57.759 | 1.00th=[31327], 5.00th=[32375], 10.00th=[32375], 20.00th=[32637], 00:37:57.759 | 30.00th=[32637], 40.00th=[32900], 50.00th=[32900], 60.00th=[32900], 00:37:57.759 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.759 | 99.00th=[34341], 99.50th=[34341], 99.90th=[61080], 99.95th=[61080], 00:37:57.759 | 99.99th=[61080] 00:37:57.759 bw ( KiB/s): min= 1795, max= 2048, per=4.13%, avg=1920.16, stdev=42.17, samples=19 00:37:57.759 iops : min= 448, max= 512, avg=480.00, stdev=10.67, samples=19 00:37:57.759 lat (msec) : 20=0.33%, 50=99.34%, 100=0.33% 00:37:57.759 cpu : usr=99.05%, sys=0.67%, ctx=35, majf=0, minf=44 00:37:57.759 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=4816,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename0: (groupid=0, jobs=1): err= 0: pid=4192356: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=502, BW=2009KiB/s (2057kB/s)(19.6MiB/10017msec) 00:37:57.759 slat (nsec): min=2951, max=63563, avg=14204.40, stdev=10287.17 00:37:57.759 clat (usec): min=1004, max=48389, avg=31735.39, stdev=6135.53 00:37:57.759 lat (usec): min=1009, max=48404, avg=31749.59, stdev=6137.30 00:37:57.759 clat percentiles (usec): 00:37:57.759 | 1.00th=[ 1516], 5.00th=[19792], 10.00th=[32637], 20.00th=[32900], 00:37:57.759 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.759 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.759 | 99.00th=[34866], 99.50th=[35390], 99.90th=[48497], 99.95th=[48497], 00:37:57.759 | 99.99th=[48497] 00:37:57.759 bw ( KiB/s): min= 1904, max= 3497, per=4.32%, avg=2010.85, stdev=351.78, samples=20 00:37:57.759 iops : min= 476, max= 874, avg=502.70, stdev=87.89, samples=20 00:37:57.759 lat (msec) : 2=2.58%, 4=0.40%, 10=0.62%, 20=1.43%, 50=94.97% 00:37:57.759 cpu : usr=99.03%, sys=0.72%, ctx=13, majf=0, minf=50 00:37:57.759 IO depths : 1=1.2%, 2=7.2%, 4=24.1%, 8=56.0%, 16=11.5%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=94.2%, 8=0.4%, 16=5.4%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=5030,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename0: (groupid=0, jobs=1): err= 0: pid=4192357: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=488, BW=1954KiB/s (2001kB/s)(19.1MiB/10016msec) 00:37:57.759 slat (nsec): min=5670, max=65808, avg=10066.86, stdev=7745.91 00:37:57.759 clat (usec): min=9683, max=47505, avg=32677.10, stdev=2796.08 00:37:57.759 lat (usec): min=9695, max=47511, avg=32687.17, stdev=2795.67 00:37:57.759 clat percentiles (usec): 00:37:57.759 | 1.00th=[15139], 5.00th=[32375], 10.00th=[32637], 20.00th=[32900], 00:37:57.759 | 30.00th=[32900], 40.00th=[32900], 50.00th=[33162], 60.00th=[33162], 00:37:57.759 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.759 | 99.00th=[34341], 99.50th=[34866], 99.90th=[41681], 99.95th=[41681], 00:37:57.759 | 99.99th=[47449] 00:37:57.759 bw ( KiB/s): min= 1792, max= 2400, per=4.19%, avg=1950.40, stdev=117.22, samples=20 00:37:57.759 iops : min= 448, max= 600, avg=487.60, stdev=29.30, samples=20 00:37:57.759 lat (msec) : 10=0.14%, 20=1.78%, 50=98.08% 00:37:57.759 cpu : usr=99.06%, sys=0.68%, ctx=13, majf=0, minf=36 00:37:57.759 IO depths : 1=6.0%, 2=12.1%, 4=24.4%, 8=51.0%, 16=6.5%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=94.0%, 8=0.2%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=4892,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename0: (groupid=0, jobs=1): err= 0: pid=4192358: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=482, BW=1930KiB/s (1976kB/s)(18.9MiB/10014msec) 00:37:57.759 slat (nsec): min=5728, max=73869, avg=18218.95, stdev=12661.59 00:37:57.759 clat (usec): min=18641, max=47476, avg=32978.29, stdev=1006.21 00:37:57.759 lat (usec): min=18648, max=47483, avg=32996.51, stdev=1006.83 00:37:57.759 clat percentiles (usec): 00:37:57.759 | 1.00th=[31327], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.759 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.759 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33424], 95.00th=[33817], 00:37:57.759 | 99.00th=[34341], 99.50th=[34866], 99.90th=[35390], 99.95th=[35390], 00:37:57.759 | 99.99th=[47449] 00:37:57.759 bw ( KiB/s): min= 1920, max= 2048, per=4.14%, avg=1926.74, stdev=29.37, samples=19 00:37:57.759 iops : min= 480, max= 512, avg=481.68, stdev= 7.34, samples=19 00:37:57.759 lat (msec) : 20=0.04%, 50=99.96% 00:37:57.759 cpu : usr=98.54%, sys=1.04%, ctx=85, majf=0, minf=47 00:37:57.759 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename0: (groupid=0, jobs=1): err= 0: pid=4192359: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=485, BW=1943KiB/s (1989kB/s)(19.0MiB/10015msec) 00:37:57.759 slat (nsec): min=5777, max=70261, avg=16621.88, stdev=10597.18 00:37:57.759 clat (usec): min=10997, max=44760, avg=32797.35, stdev=2142.48 00:37:57.759 lat (usec): min=11008, max=44768, avg=32813.98, stdev=2143.17 00:37:57.759 clat percentiles (usec): 00:37:57.759 | 1.00th=[22152], 5.00th=[32375], 10.00th=[32637], 20.00th=[32900], 00:37:57.759 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.759 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33424], 95.00th=[33817], 00:37:57.759 | 99.00th=[34341], 99.50th=[34866], 99.90th=[34866], 99.95th=[42730], 00:37:57.759 | 99.99th=[44827] 00:37:57.759 bw ( KiB/s): min= 1920, max= 2048, per=4.17%, avg=1939.20, stdev=46.89, samples=20 00:37:57.759 iops : min= 480, max= 512, avg=484.80, stdev=11.72, samples=20 00:37:57.759 lat (msec) : 20=0.99%, 50=99.01% 00:37:57.759 cpu : usr=98.86%, sys=0.88%, ctx=12, majf=0, minf=44 00:37:57.759 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.759 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.759 issued rwts: total=4864,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.759 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.759 filename1: (groupid=0, jobs=1): err= 0: pid=4192360: Tue Nov 26 19:34:09 2024 00:37:57.759 read: IOPS=484, BW=1938KiB/s (1984kB/s)(18.9MiB/10007msec) 00:37:57.760 slat (nsec): min=5711, max=90719, avg=22141.77, stdev=14780.39 00:37:57.760 clat (usec): min=11070, max=34766, avg=32831.05, stdev=1871.51 00:37:57.760 lat (usec): min=11079, max=34782, avg=32853.20, stdev=1872.07 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[23987], 5.00th=[32375], 10.00th=[32375], 20.00th=[32637], 00:37:57.760 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[32900], 00:37:57.760 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.760 | 99.00th=[34341], 99.50th=[34341], 99.90th=[34866], 99.95th=[34866], 00:37:57.760 | 99.99th=[34866] 00:37:57.760 bw ( KiB/s): min= 1920, max= 2048, per=4.17%, avg=1940.21, stdev=47.95, samples=19 00:37:57.760 iops : min= 480, max= 512, avg=485.05, stdev=11.99, samples=19 00:37:57.760 lat (msec) : 20=0.66%, 50=99.34% 00:37:57.760 cpu : usr=98.82%, sys=0.91%, ctx=11, majf=0, minf=29 00:37:57.760 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.760 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 issued rwts: total=4848,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.760 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.760 filename1: (groupid=0, jobs=1): err= 0: pid=4192361: Tue Nov 26 19:34:09 2024 00:37:57.760 read: IOPS=483, BW=1932KiB/s (1979kB/s)(18.9MiB/10015msec) 00:37:57.760 slat (nsec): min=5675, max=96650, avg=19400.31, stdev=15767.21 00:37:57.760 clat (usec): min=13938, max=44557, avg=32955.10, stdev=1440.94 00:37:57.760 lat (usec): min=13945, max=44612, avg=32974.50, stdev=1440.88 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[24249], 5.00th=[32375], 10.00th=[32375], 20.00th=[32637], 00:37:57.760 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.760 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.760 | 99.00th=[34341], 99.50th=[38536], 99.90th=[40109], 99.95th=[40109], 00:37:57.760 | 99.99th=[44303] 00:37:57.760 bw ( KiB/s): min= 1792, max= 2048, per=4.14%, avg=1929.26, stdev=52.61, samples=19 00:37:57.760 iops : min= 448, max= 512, avg=482.32, stdev=13.15, samples=19 00:37:57.760 lat (msec) : 20=0.12%, 50=99.88% 00:37:57.760 cpu : usr=98.74%, sys=0.92%, ctx=105, majf=0, minf=38 00:37:57.760 IO depths : 1=6.1%, 2=12.3%, 4=24.7%, 8=50.5%, 16=6.4%, 32=0.0%, >=64=0.0% 00:37:57.760 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 complete : 0=0.0%, 4=94.0%, 8=0.1%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 issued rwts: total=4838,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.760 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.760 filename1: (groupid=0, jobs=1): err= 0: pid=4192362: Tue Nov 26 19:34:09 2024 00:37:57.760 read: IOPS=482, BW=1931KiB/s (1978kB/s)(18.9MiB/10007msec) 00:37:57.760 slat (nsec): min=5655, max=85873, avg=15712.78, stdev=12648.16 00:37:57.760 clat (usec): min=13875, max=91234, avg=33060.18, stdev=5509.67 00:37:57.760 lat (usec): min=13897, max=91250, avg=33075.90, stdev=5509.84 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[19792], 5.00th=[24773], 10.00th=[26870], 20.00th=[31851], 00:37:57.760 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.760 | 70.00th=[33424], 80.00th=[33817], 90.00th=[38011], 95.00th=[41681], 00:37:57.760 | 99.00th=[51119], 99.50th=[54264], 99.90th=[90702], 99.95th=[90702], 00:37:57.760 | 99.99th=[91751] 00:37:57.760 bw ( KiB/s): min= 1632, max= 2064, per=4.14%, avg=1928.42, stdev=101.63, samples=19 00:37:57.760 iops : min= 408, max= 516, avg=482.11, stdev=25.41, samples=19 00:37:57.760 lat (msec) : 20=1.35%, 50=97.48%, 100=1.18% 00:37:57.760 cpu : usr=98.95%, sys=0.78%, ctx=12, majf=0, minf=28 00:37:57.760 IO depths : 1=0.1%, 2=0.5%, 4=4.3%, 8=79.0%, 16=16.1%, 32=0.0%, >=64=0.0% 00:37:57.760 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 complete : 0=0.0%, 4=89.5%, 8=8.3%, 16=2.2%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.760 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.760 filename1: (groupid=0, jobs=1): err= 0: pid=4192363: Tue Nov 26 19:34:09 2024 00:37:57.760 read: IOPS=486, BW=1947KiB/s (1994kB/s)(19.1MiB/10024msec) 00:37:57.760 slat (nsec): min=5673, max=66900, avg=8148.56, stdev=4685.72 00:37:57.760 clat (usec): min=10866, max=44807, avg=32793.52, stdev=2293.31 00:37:57.760 lat (usec): min=10886, max=44820, avg=32801.67, stdev=2292.19 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[19268], 5.00th=[32375], 10.00th=[32637], 20.00th=[32900], 00:37:57.760 | 30.00th=[32900], 40.00th=[32900], 50.00th=[33162], 60.00th=[33162], 00:37:57.760 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.760 | 99.00th=[34341], 99.50th=[34866], 99.90th=[34866], 99.95th=[34866], 00:37:57.760 | 99.99th=[44827] 00:37:57.760 bw ( KiB/s): min= 1920, max= 2048, per=4.18%, avg=1945.60, stdev=52.53, samples=20 00:37:57.760 iops : min= 480, max= 512, avg=486.40, stdev=13.13, samples=20 00:37:57.760 lat (msec) : 20=1.52%, 50=98.48% 00:37:57.760 cpu : usr=98.59%, sys=0.98%, ctx=63, majf=0, minf=30 00:37:57.760 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.760 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 issued rwts: total=4880,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.760 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.760 filename1: (groupid=0, jobs=1): err= 0: pid=4192364: Tue Nov 26 19:34:09 2024 00:37:57.760 read: IOPS=481, BW=1925KiB/s (1971kB/s)(18.8MiB/10006msec) 00:37:57.760 slat (nsec): min=5510, max=89467, avg=22943.39, stdev=15904.67 00:37:57.760 clat (usec): min=18730, max=60782, avg=33049.26, stdev=1923.15 00:37:57.760 lat (usec): min=18736, max=60798, avg=33072.20, stdev=1922.36 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[31065], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.760 | 30.00th=[32637], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.760 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.760 | 99.00th=[34341], 99.50th=[34866], 99.90th=[60556], 99.95th=[60556], 00:37:57.760 | 99.99th=[60556] 00:37:57.760 bw ( KiB/s): min= 1795, max= 2048, per=4.13%, avg=1920.16, stdev=42.17, samples=19 00:37:57.760 iops : min= 448, max= 512, avg=480.00, stdev=10.67, samples=19 00:37:57.760 lat (msec) : 20=0.33%, 50=99.34%, 100=0.33% 00:37:57.760 cpu : usr=98.57%, sys=0.98%, ctx=40, majf=0, minf=27 00:37:57.760 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.760 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 issued rwts: total=4816,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.760 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.760 filename1: (groupid=0, jobs=1): err= 0: pid=4192365: Tue Nov 26 19:34:09 2024 00:37:57.760 read: IOPS=498, BW=1995KiB/s (2043kB/s)(19.5MiB/10011msec) 00:37:57.760 slat (nsec): min=5669, max=97394, avg=16684.33, stdev=12386.08 00:37:57.760 clat (usec): min=13266, max=56988, avg=31946.31, stdev=4534.68 00:37:57.760 lat (usec): min=13289, max=57006, avg=31963.00, stdev=4536.02 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[19530], 5.00th=[21627], 10.00th=[25822], 20.00th=[30278], 00:37:57.760 | 30.00th=[32637], 40.00th=[32900], 50.00th=[32900], 60.00th=[32900], 00:37:57.760 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[38011], 00:37:57.760 | 99.00th=[43254], 99.50th=[46400], 99.90th=[56886], 99.95th=[56886], 00:37:57.760 | 99.99th=[56886] 00:37:57.760 bw ( KiB/s): min= 1792, max= 2160, per=4.27%, avg=1988.21, stdev=101.73, samples=19 00:37:57.760 iops : min= 448, max= 540, avg=497.05, stdev=25.43, samples=19 00:37:57.760 lat (msec) : 20=1.54%, 50=98.14%, 100=0.32% 00:37:57.760 cpu : usr=98.50%, sys=0.99%, ctx=145, majf=0, minf=32 00:37:57.760 IO depths : 1=3.2%, 2=6.5%, 4=14.9%, 8=64.8%, 16=10.7%, 32=0.0%, >=64=0.0% 00:37:57.760 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 complete : 0=0.0%, 4=91.6%, 8=4.1%, 16=4.3%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 issued rwts: total=4994,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.760 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.760 filename1: (groupid=0, jobs=1): err= 0: pid=4192366: Tue Nov 26 19:34:09 2024 00:37:57.760 read: IOPS=481, BW=1925KiB/s (1971kB/s)(18.8MiB/10008msec) 00:37:57.760 slat (nsec): min=5748, max=92712, avg=24314.03, stdev=15779.61 00:37:57.760 clat (usec): min=13146, max=61378, avg=33001.29, stdev=2176.59 00:37:57.760 lat (usec): min=13169, max=61395, avg=33025.61, stdev=2176.64 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[31327], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.760 | 30.00th=[32637], 40.00th=[32900], 50.00th=[32900], 60.00th=[32900], 00:37:57.760 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.760 | 99.00th=[34866], 99.50th=[40633], 99.90th=[61604], 99.95th=[61604], 00:37:57.760 | 99.99th=[61604] 00:37:57.760 bw ( KiB/s): min= 1792, max= 2048, per=4.13%, avg=1920.16, stdev=59.99, samples=19 00:37:57.760 iops : min= 448, max= 512, avg=480.00, stdev=15.08, samples=19 00:37:57.760 lat (msec) : 20=0.33%, 50=99.34%, 100=0.33% 00:37:57.760 cpu : usr=98.76%, sys=0.79%, ctx=58, majf=0, minf=32 00:37:57.760 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:37:57.760 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.760 issued rwts: total=4816,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.760 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.760 filename1: (groupid=0, jobs=1): err= 0: pid=4192367: Tue Nov 26 19:34:09 2024 00:37:57.760 read: IOPS=482, BW=1930KiB/s (1976kB/s)(18.9MiB/10014msec) 00:37:57.760 slat (nsec): min=6086, max=63383, avg=19467.75, stdev=11358.45 00:37:57.760 clat (usec): min=18783, max=47154, avg=32967.37, stdev=1005.06 00:37:57.760 lat (usec): min=18792, max=47168, avg=32986.84, stdev=1005.79 00:37:57.760 clat percentiles (usec): 00:37:57.760 | 1.00th=[31327], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.760 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[32900], 00:37:57.760 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33424], 95.00th=[33817], 00:37:57.760 | 99.00th=[34341], 99.50th=[34866], 99.90th=[35390], 99.95th=[35390], 00:37:57.760 | 99.99th=[46924] 00:37:57.760 bw ( KiB/s): min= 1920, max= 2048, per=4.14%, avg=1926.74, stdev=29.37, samples=19 00:37:57.760 iops : min= 480, max= 512, avg=481.68, stdev= 7.34, samples=19 00:37:57.761 lat (msec) : 20=0.04%, 50=99.96% 00:37:57.761 cpu : usr=98.44%, sys=1.11%, ctx=42, majf=0, minf=50 00:37:57.761 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.761 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.761 filename2: (groupid=0, jobs=1): err= 0: pid=4192368: Tue Nov 26 19:34:09 2024 00:37:57.761 read: IOPS=482, BW=1931KiB/s (1978kB/s)(18.9MiB/10008msec) 00:37:57.761 slat (nsec): min=5693, max=57709, avg=11303.87, stdev=8104.09 00:37:57.761 clat (usec): min=18024, max=47176, avg=33042.21, stdev=1186.73 00:37:57.761 lat (usec): min=18032, max=47191, avg=33053.52, stdev=1186.39 00:37:57.761 clat percentiles (usec): 00:37:57.761 | 1.00th=[30802], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.761 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.761 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.761 | 99.00th=[34866], 99.50th=[34866], 99.90th=[44827], 99.95th=[45351], 00:37:57.761 | 99.99th=[46924] 00:37:57.761 bw ( KiB/s): min= 1920, max= 2032, per=4.14%, avg=1926.74, stdev=25.75, samples=19 00:37:57.761 iops : min= 480, max= 508, avg=481.68, stdev= 6.44, samples=19 00:37:57.761 lat (msec) : 20=0.08%, 50=99.92% 00:37:57.761 cpu : usr=99.06%, sys=0.67%, ctx=6, majf=0, minf=43 00:37:57.761 IO depths : 1=6.0%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.5%, 32=0.0%, >=64=0.0% 00:37:57.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.761 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.761 filename2: (groupid=0, jobs=1): err= 0: pid=4192369: Tue Nov 26 19:34:09 2024 00:37:57.761 read: IOPS=482, BW=1930KiB/s (1976kB/s)(18.9MiB/10016msec) 00:37:57.761 slat (nsec): min=5137, max=71346, avg=11734.45, stdev=9920.29 00:37:57.761 clat (usec): min=18716, max=44511, avg=33073.51, stdev=1350.64 00:37:57.761 lat (usec): min=18729, max=44539, avg=33085.25, stdev=1350.05 00:37:57.761 clat percentiles (usec): 00:37:57.761 | 1.00th=[25822], 5.00th=[32375], 10.00th=[32637], 20.00th=[32900], 00:37:57.761 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.761 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:37:57.761 | 99.00th=[34866], 99.50th=[39584], 99.90th=[41157], 99.95th=[41681], 00:37:57.761 | 99.99th=[44303] 00:37:57.761 bw ( KiB/s): min= 1792, max= 2048, per=4.14%, avg=1926.74, stdev=51.80, samples=19 00:37:57.761 iops : min= 448, max= 512, avg=481.68, stdev=12.95, samples=19 00:37:57.761 lat (msec) : 20=0.33%, 50=99.67% 00:37:57.761 cpu : usr=98.99%, sys=0.75%, ctx=17, majf=0, minf=59 00:37:57.761 IO depths : 1=5.8%, 2=12.0%, 4=25.0%, 8=50.5%, 16=6.7%, 32=0.0%, >=64=0.0% 00:37:57.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.761 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.761 filename2: (groupid=0, jobs=1): err= 0: pid=4192370: Tue Nov 26 19:34:09 2024 00:37:57.761 read: IOPS=485, BW=1941KiB/s (1987kB/s)(19.0MiB/10008msec) 00:37:57.761 slat (nsec): min=5660, max=85843, avg=19887.04, stdev=14340.96 00:37:57.761 clat (usec): min=16430, max=61813, avg=32807.57, stdev=3364.11 00:37:57.761 lat (usec): min=16436, max=61830, avg=32827.45, stdev=3364.92 00:37:57.761 clat percentiles (usec): 00:37:57.761 | 1.00th=[21890], 5.00th=[26608], 10.00th=[29754], 20.00th=[32637], 00:37:57.761 | 30.00th=[32637], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.761 | 70.00th=[33162], 80.00th=[33817], 90.00th=[34341], 95.00th=[36963], 00:37:57.761 | 99.00th=[42206], 99.50th=[44303], 99.90th=[61604], 99.95th=[61604], 00:37:57.761 | 99.99th=[61604] 00:37:57.761 bw ( KiB/s): min= 1792, max= 2032, per=4.15%, avg=1933.47, stdev=53.14, samples=19 00:37:57.761 iops : min= 448, max= 508, avg=483.37, stdev=13.28, samples=19 00:37:57.761 lat (msec) : 20=0.62%, 50=99.05%, 100=0.33% 00:37:57.761 cpu : usr=99.02%, sys=0.71%, ctx=12, majf=0, minf=46 00:37:57.761 IO depths : 1=3.7%, 2=7.4%, 4=15.8%, 8=62.7%, 16=10.4%, 32=0.0%, >=64=0.0% 00:37:57.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 complete : 0=0.0%, 4=91.9%, 8=4.0%, 16=4.2%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 issued rwts: total=4856,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.761 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.761 filename2: (groupid=0, jobs=1): err= 0: pid=4192371: Tue Nov 26 19:34:09 2024 00:37:57.761 read: IOPS=482, BW=1929KiB/s (1975kB/s)(18.9MiB/10020msec) 00:37:57.761 slat (nsec): min=5694, max=70400, avg=15042.89, stdev=10164.75 00:37:57.761 clat (usec): min=22227, max=38773, avg=33046.41, stdev=864.96 00:37:57.761 lat (usec): min=22237, max=38791, avg=33061.45, stdev=864.38 00:37:57.761 clat percentiles (usec): 00:37:57.761 | 1.00th=[31589], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.761 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.761 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.761 | 99.00th=[34341], 99.50th=[34866], 99.90th=[38536], 99.95th=[38536], 00:37:57.761 | 99.99th=[38536] 00:37:57.761 bw ( KiB/s): min= 1792, max= 2048, per=4.14%, avg=1926.74, stdev=51.80, samples=19 00:37:57.761 iops : min= 448, max= 512, avg=481.68, stdev=12.95, samples=19 00:37:57.761 lat (msec) : 50=100.00% 00:37:57.761 cpu : usr=98.82%, sys=0.92%, ctx=16, majf=0, minf=35 00:37:57.761 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:37:57.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.761 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.761 filename2: (groupid=0, jobs=1): err= 0: pid=4192372: Tue Nov 26 19:34:09 2024 00:37:57.761 read: IOPS=497, BW=1991KiB/s (2039kB/s)(19.5MiB/10016msec) 00:37:57.761 slat (usec): min=5, max=100, avg=16.67, stdev=13.87 00:37:57.761 clat (usec): min=7457, max=52047, avg=31994.82, stdev=4334.19 00:37:57.761 lat (usec): min=7473, max=52079, avg=32011.49, stdev=4336.32 00:37:57.761 clat percentiles (usec): 00:37:57.761 | 1.00th=[19792], 5.00th=[22938], 10.00th=[25822], 20.00th=[32375], 00:37:57.761 | 30.00th=[32637], 40.00th=[32900], 50.00th=[32900], 60.00th=[32900], 00:37:57.761 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[35914], 00:37:57.761 | 99.00th=[45876], 99.50th=[46924], 99.90th=[52167], 99.95th=[52167], 00:37:57.761 | 99.99th=[52167] 00:37:57.761 bw ( KiB/s): min= 1888, max= 2240, per=4.28%, avg=1990.40, stdev=114.25, samples=20 00:37:57.761 iops : min= 472, max= 560, avg=497.60, stdev=28.56, samples=20 00:37:57.761 lat (msec) : 10=0.06%, 20=1.48%, 50=98.34%, 100=0.12% 00:37:57.761 cpu : usr=98.75%, sys=0.92%, ctx=43, majf=0, minf=52 00:37:57.761 IO depths : 1=4.4%, 2=8.8%, 4=19.3%, 8=59.0%, 16=8.4%, 32=0.0%, >=64=0.0% 00:37:57.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 complete : 0=0.0%, 4=92.6%, 8=1.9%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.761 issued rwts: total=4986,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.761 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.761 filename2: (groupid=0, jobs=1): err= 0: pid=4192373: Tue Nov 26 19:34:09 2024 00:37:57.762 read: IOPS=481, BW=1925KiB/s (1971kB/s)(18.8MiB/10007msec) 00:37:57.762 slat (nsec): min=5798, max=84493, avg=23035.05, stdev=13747.98 00:37:57.762 clat (usec): min=18742, max=61252, avg=33035.19, stdev=1965.07 00:37:57.762 lat (usec): min=18760, max=61268, avg=33058.22, stdev=1964.63 00:37:57.762 clat percentiles (usec): 00:37:57.762 | 1.00th=[31065], 5.00th=[32375], 10.00th=[32637], 20.00th=[32637], 00:37:57.762 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[32900], 00:37:57.762 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.762 | 99.00th=[34341], 99.50th=[34866], 99.90th=[61080], 99.95th=[61080], 00:37:57.762 | 99.99th=[61080] 00:37:57.762 bw ( KiB/s): min= 1795, max= 2048, per=4.13%, avg=1920.16, stdev=42.17, samples=19 00:37:57.762 iops : min= 448, max= 512, avg=480.00, stdev=10.67, samples=19 00:37:57.762 lat (msec) : 20=0.33%, 50=99.34%, 100=0.33% 00:37:57.762 cpu : usr=98.83%, sys=0.91%, ctx=20, majf=0, minf=37 00:37:57.762 IO depths : 1=6.2%, 2=12.5%, 4=24.9%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:37:57.762 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.762 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.762 issued rwts: total=4816,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.762 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.762 filename2: (groupid=0, jobs=1): err= 0: pid=4192374: Tue Nov 26 19:34:09 2024 00:37:57.762 read: IOPS=480, BW=1921KiB/s (1967kB/s)(18.8MiB/10007msec) 00:37:57.762 slat (nsec): min=5506, max=89501, avg=15096.66, stdev=11810.94 00:37:57.762 clat (usec): min=9402, max=60895, avg=33246.79, stdev=2984.58 00:37:57.762 lat (usec): min=9408, max=60911, avg=33261.88, stdev=2984.45 00:37:57.762 clat percentiles (usec): 00:37:57.762 | 1.00th=[22414], 5.00th=[32375], 10.00th=[32637], 20.00th=[32900], 00:37:57.762 | 30.00th=[32900], 40.00th=[32900], 50.00th=[33162], 60.00th=[33162], 00:37:57.762 | 70.00th=[33424], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:37:57.762 | 99.00th=[47449], 99.50th=[49021], 99.90th=[61080], 99.95th=[61080], 00:37:57.762 | 99.99th=[61080] 00:37:57.762 bw ( KiB/s): min= 1619, max= 2048, per=4.11%, avg=1913.42, stdev=85.14, samples=19 00:37:57.762 iops : min= 404, max= 512, avg=478.32, stdev=21.43, samples=19 00:37:57.762 lat (msec) : 10=0.12%, 20=0.71%, 50=98.71%, 100=0.46% 00:37:57.762 cpu : usr=99.22%, sys=0.53%, ctx=11, majf=0, minf=32 00:37:57.762 IO depths : 1=0.8%, 2=1.7%, 4=4.1%, 8=76.6%, 16=16.7%, 32=0.0%, >=64=0.0% 00:37:57.762 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.762 complete : 0=0.0%, 4=90.1%, 8=8.9%, 16=1.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.762 issued rwts: total=4806,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.762 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.762 filename2: (groupid=0, jobs=1): err= 0: pid=4192375: Tue Nov 26 19:34:09 2024 00:37:57.762 read: IOPS=485, BW=1942KiB/s (1989kB/s)(19.0MiB/10016msec) 00:37:57.762 slat (nsec): min=5741, max=76792, avg=15706.82, stdev=10963.29 00:37:57.762 clat (usec): min=10999, max=41984, avg=32822.13, stdev=2258.40 00:37:57.762 lat (usec): min=11016, max=41992, avg=32837.84, stdev=2258.56 00:37:57.762 clat percentiles (usec): 00:37:57.762 | 1.00th=[19006], 5.00th=[32375], 10.00th=[32637], 20.00th=[32900], 00:37:57.762 | 30.00th=[32900], 40.00th=[32900], 50.00th=[32900], 60.00th=[33162], 00:37:57.762 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[33817], 00:37:57.762 | 99.00th=[34866], 99.50th=[34866], 99.90th=[41157], 99.95th=[41157], 00:37:57.762 | 99.99th=[42206] 00:37:57.762 bw ( KiB/s): min= 1920, max= 2176, per=4.17%, avg=1939.20, stdev=62.64, samples=20 00:37:57.762 iops : min= 480, max= 544, avg=484.80, stdev=15.66, samples=20 00:37:57.762 lat (msec) : 20=1.36%, 50=98.64% 00:37:57.762 cpu : usr=98.78%, sys=0.77%, ctx=86, majf=0, minf=42 00:37:57.762 IO depths : 1=6.0%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.5%, 32=0.0%, >=64=0.0% 00:37:57.762 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.762 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.762 issued rwts: total=4864,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.762 latency : target=0, window=0, percentile=100.00%, depth=16 00:37:57.762 00:37:57.762 Run status group 0 (all jobs): 00:37:57.762 READ: bw=45.5MiB/s (47.7MB/s), 1921KiB/s-2009KiB/s (1967kB/s-2057kB/s), io=456MiB (478MB), run=10006-10024msec 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@113 -- # destroy_subsystems 0 1 2 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 2 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=2 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null2 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # NULL_DIF=1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # bs=8k,16k,128k 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # numjobs=2 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # iodepth=8 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # runtime=5 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # files=1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@117 -- # create_subsystems 0 1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 bdev_null0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 [2024-11-26 19:34:09.271465] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.762 bdev_null1 00:37:57.762 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # fio /dev/fd/62 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # create_json_sub_conf 0 1 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # config=() 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # local subsystem config 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:57.763 { 00:37:57.763 "params": { 00:37:57.763 "name": "Nvme$subsystem", 00:37:57.763 "trtype": "$TEST_TRANSPORT", 00:37:57.763 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:57.763 "adrfam": "ipv4", 00:37:57.763 "trsvcid": "$NVMF_PORT", 00:37:57.763 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:57.763 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:57.763 "hdgst": ${hdgst:-false}, 00:37:57.763 "ddgst": ${ddgst:-false} 00:37:57.763 }, 00:37:57.763 "method": "bdev_nvme_attach_controller" 00:37:57.763 } 00:37:57.763 EOF 00:37:57.763 )") 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local sanitizers 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # shift 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1347 -- # local asan_lib= 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libasan 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:37:57.763 { 00:37:57.763 "params": { 00:37:57.763 "name": "Nvme$subsystem", 00:37:57.763 "trtype": "$TEST_TRANSPORT", 00:37:57.763 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:57.763 "adrfam": "ipv4", 00:37:57.763 "trsvcid": "$NVMF_PORT", 00:37:57.763 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:57.763 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:57.763 "hdgst": ${hdgst:-false}, 00:37:57.763 "ddgst": ${ddgst:-false} 00:37:57.763 }, 00:37:57.763 "method": "bdev_nvme_attach_controller" 00:37:57.763 } 00:37:57.763 EOF 00:37:57.763 )") 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # jq . 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@585 -- # IFS=, 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:37:57.763 "params": { 00:37:57.763 "name": "Nvme0", 00:37:57.763 "trtype": "tcp", 00:37:57.763 "traddr": "10.0.0.2", 00:37:57.763 "adrfam": "ipv4", 00:37:57.763 "trsvcid": "4420", 00:37:57.763 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:57.763 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:57.763 "hdgst": false, 00:37:57.763 "ddgst": false 00:37:57.763 }, 00:37:57.763 "method": "bdev_nvme_attach_controller" 00:37:57.763 },{ 00:37:57.763 "params": { 00:37:57.763 "name": "Nvme1", 00:37:57.763 "trtype": "tcp", 00:37:57.763 "traddr": "10.0.0.2", 00:37:57.763 "adrfam": "ipv4", 00:37:57.763 "trsvcid": "4420", 00:37:57.763 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:37:57.763 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:37:57.763 "hdgst": false, 00:37:57.763 "ddgst": false 00:37:57.763 }, 00:37:57.763 "method": "bdev_nvme_attach_controller" 00:37:57.763 }' 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:37:57.763 19:34:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:37:57.763 filename0: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:37:57.763 ... 00:37:57.763 filename1: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:37:57.763 ... 00:37:57.763 fio-3.35 00:37:57.763 Starting 4 threads 00:38:03.053 00:38:03.053 filename0: (groupid=0, jobs=1): err= 0: pid=1030: Tue Nov 26 19:34:15 2024 00:38:03.053 read: IOPS=2218, BW=17.3MiB/s (18.2MB/s)(86.7MiB/5003msec) 00:38:03.053 slat (nsec): min=5504, max=86556, avg=6424.95, stdev=2943.70 00:38:03.053 clat (usec): min=1427, max=6166, avg=3588.73, stdev=628.13 00:38:03.053 lat (usec): min=1448, max=6172, avg=3595.16, stdev=628.09 00:38:03.053 clat percentiles (usec): 00:38:03.053 | 1.00th=[ 2474], 5.00th=[ 2737], 10.00th=[ 2933], 20.00th=[ 3163], 00:38:03.053 | 30.00th=[ 3261], 40.00th=[ 3392], 50.00th=[ 3490], 60.00th=[ 3589], 00:38:03.053 | 70.00th=[ 3687], 80.00th=[ 3916], 90.00th=[ 4621], 95.00th=[ 5014], 00:38:03.053 | 99.00th=[ 5407], 99.50th=[ 5407], 99.90th=[ 5932], 99.95th=[ 5997], 00:38:03.053 | 99.99th=[ 6194] 00:38:03.053 bw ( KiB/s): min=17552, max=18032, per=27.35%, avg=17788.44, stdev=174.68, samples=9 00:38:03.053 iops : min= 2194, max= 2254, avg=2223.56, stdev=21.84, samples=9 00:38:03.053 lat (msec) : 2=0.17%, 4=83.63%, 10=16.20% 00:38:03.053 cpu : usr=96.58%, sys=3.12%, ctx=48, majf=0, minf=9 00:38:03.053 IO depths : 1=0.1%, 2=0.5%, 4=70.6%, 8=28.8%, 16=0.0%, 32=0.0%, >=64=0.0% 00:38:03.053 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 complete : 0=0.0%, 4=93.9%, 8=6.1%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 issued rwts: total=11099,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:38:03.053 latency : target=0, window=0, percentile=100.00%, depth=8 00:38:03.053 filename0: (groupid=0, jobs=1): err= 0: pid=1031: Tue Nov 26 19:34:15 2024 00:38:03.053 read: IOPS=1975, BW=15.4MiB/s (16.2MB/s)(77.2MiB/5002msec) 00:38:03.053 slat (nsec): min=5486, max=65172, avg=6156.74, stdev=2306.79 00:38:03.053 clat (usec): min=1565, max=10113, avg=4031.97, stdev=731.32 00:38:03.053 lat (usec): min=1570, max=10151, avg=4038.12, stdev=731.33 00:38:03.053 clat percentiles (usec): 00:38:03.053 | 1.00th=[ 3064], 5.00th=[ 3326], 10.00th=[ 3392], 20.00th=[ 3589], 00:38:03.053 | 30.00th=[ 3654], 40.00th=[ 3752], 50.00th=[ 3851], 60.00th=[ 3916], 00:38:03.053 | 70.00th=[ 4047], 80.00th=[ 4228], 90.00th=[ 5407], 95.00th=[ 5735], 00:38:03.053 | 99.00th=[ 5997], 99.50th=[ 6325], 99.90th=[ 6849], 99.95th=[ 9765], 00:38:03.053 | 99.99th=[10159] 00:38:03.053 bw ( KiB/s): min=15185, max=16480, per=24.25%, avg=15770.78, stdev=347.67, samples=9 00:38:03.053 iops : min= 1898, max= 2060, avg=1971.33, stdev=43.49, samples=9 00:38:03.053 lat (msec) : 2=0.22%, 4=67.02%, 10=32.74%, 20=0.01% 00:38:03.053 cpu : usr=97.32%, sys=2.46%, ctx=7, majf=0, minf=9 00:38:03.053 IO depths : 1=0.1%, 2=0.1%, 4=73.0%, 8=26.9%, 16=0.0%, 32=0.0%, >=64=0.0% 00:38:03.053 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 complete : 0=0.0%, 4=92.3%, 8=7.7%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 issued rwts: total=9883,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:38:03.053 latency : target=0, window=0, percentile=100.00%, depth=8 00:38:03.053 filename1: (groupid=0, jobs=1): err= 0: pid=1032: Tue Nov 26 19:34:15 2024 00:38:03.053 read: IOPS=1952, BW=15.3MiB/s (16.0MB/s)(76.3MiB/5002msec) 00:38:03.053 slat (nsec): min=5483, max=64182, avg=6103.50, stdev=2139.39 00:38:03.053 clat (usec): min=1649, max=6909, avg=4080.49, stdev=744.71 00:38:03.053 lat (usec): min=1655, max=6914, avg=4086.59, stdev=744.69 00:38:03.053 clat percentiles (usec): 00:38:03.053 | 1.00th=[ 3228], 5.00th=[ 3359], 10.00th=[ 3523], 20.00th=[ 3589], 00:38:03.053 | 30.00th=[ 3654], 40.00th=[ 3752], 50.00th=[ 3851], 60.00th=[ 3916], 00:38:03.053 | 70.00th=[ 4047], 80.00th=[ 4293], 90.00th=[ 5473], 95.00th=[ 5735], 00:38:03.053 | 99.00th=[ 6128], 99.50th=[ 6259], 99.90th=[ 6587], 99.95th=[ 6652], 00:38:03.053 | 99.99th=[ 6915] 00:38:03.053 bw ( KiB/s): min=15296, max=15792, per=23.95%, avg=15578.67, stdev=154.30, samples=9 00:38:03.053 iops : min= 1912, max= 1974, avg=1947.33, stdev=19.29, samples=9 00:38:03.053 lat (msec) : 2=0.15%, 4=68.17%, 10=31.68% 00:38:03.053 cpu : usr=97.00%, sys=2.76%, ctx=6, majf=0, minf=9 00:38:03.053 IO depths : 1=0.1%, 2=0.1%, 4=72.9%, 8=27.1%, 16=0.0%, 32=0.0%, >=64=0.0% 00:38:03.053 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 complete : 0=0.0%, 4=92.6%, 8=7.4%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 issued rwts: total=9767,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:38:03.053 latency : target=0, window=0, percentile=100.00%, depth=8 00:38:03.053 filename1: (groupid=0, jobs=1): err= 0: pid=1033: Tue Nov 26 19:34:15 2024 00:38:03.053 read: IOPS=1984, BW=15.5MiB/s (16.3MB/s)(77.5MiB/5001msec) 00:38:03.053 slat (nsec): min=5483, max=82864, avg=6208.90, stdev=2546.21 00:38:03.053 clat (usec): min=2110, max=6617, avg=4014.73, stdev=744.63 00:38:03.053 lat (usec): min=2115, max=6623, avg=4020.93, stdev=744.61 00:38:03.053 clat percentiles (usec): 00:38:03.053 | 1.00th=[ 2900], 5.00th=[ 3261], 10.00th=[ 3392], 20.00th=[ 3556], 00:38:03.053 | 30.00th=[ 3621], 40.00th=[ 3687], 50.00th=[ 3818], 60.00th=[ 3884], 00:38:03.053 | 70.00th=[ 4015], 80.00th=[ 4228], 90.00th=[ 5407], 95.00th=[ 5735], 00:38:03.053 | 99.00th=[ 6063], 99.50th=[ 6259], 99.90th=[ 6587], 99.95th=[ 6587], 00:38:03.053 | 99.99th=[ 6587] 00:38:03.053 bw ( KiB/s): min=15168, max=16288, per=24.39%, avg=15864.89, stdev=329.67, samples=9 00:38:03.053 iops : min= 1896, max= 2036, avg=1983.11, stdev=41.21, samples=9 00:38:03.053 lat (msec) : 4=68.52%, 10=31.48% 00:38:03.053 cpu : usr=97.02%, sys=2.74%, ctx=7, majf=0, minf=9 00:38:03.053 IO depths : 1=0.1%, 2=0.1%, 4=72.6%, 8=27.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:38:03.053 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 complete : 0=0.0%, 4=92.7%, 8=7.3%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:03.053 issued rwts: total=9924,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:38:03.053 latency : target=0, window=0, percentile=100.00%, depth=8 00:38:03.053 00:38:03.053 Run status group 0 (all jobs): 00:38:03.053 READ: bw=63.5MiB/s (66.6MB/s), 15.3MiB/s-17.3MiB/s (16.0MB/s-18.2MB/s), io=318MiB (333MB), run=5001-5003msec 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@119 -- # destroy_subsystems 0 1 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:38:03.314 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.315 00:38:03.315 real 0m24.866s 00:38:03.315 user 5m16.714s 00:38:03.315 sys 0m4.511s 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:03.315 19:34:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:38:03.315 ************************************ 00:38:03.315 END TEST fio_dif_rand_params 00:38:03.315 ************************************ 00:38:03.315 19:34:15 nvmf_dif -- target/dif.sh@144 -- # run_test fio_dif_digest fio_dif_digest 00:38:03.315 19:34:15 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:38:03.315 19:34:15 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:03.315 19:34:15 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:38:03.315 ************************************ 00:38:03.315 START TEST fio_dif_digest 00:38:03.315 ************************************ 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1129 -- # fio_dif_digest 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@123 -- # local NULL_DIF 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@124 -- # local bs numjobs runtime iodepth files 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@125 -- # local hdgst ddgst 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # NULL_DIF=3 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # bs=128k,128k,128k 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # numjobs=3 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # iodepth=3 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # runtime=10 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # hdgst=true 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # ddgst=true 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@130 -- # create_subsystems 0 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@28 -- # local sub 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@30 -- # for sub in "$@" 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@31 -- # create_subsystem 0 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@18 -- # local sub_id=0 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.315 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:38:03.577 bdev_null0 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:38:03.577 [2024-11-26 19:34:15.965120] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # fio /dev/fd/62 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # create_json_sub_conf 0 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@560 -- # config=() 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@560 -- # local subsystem config 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:38:03.577 { 00:38:03.577 "params": { 00:38:03.577 "name": "Nvme$subsystem", 00:38:03.577 "trtype": "$TEST_TRANSPORT", 00:38:03.577 "traddr": "$NVMF_FIRST_TARGET_IP", 00:38:03.577 "adrfam": "ipv4", 00:38:03.577 "trsvcid": "$NVMF_PORT", 00:38:03.577 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:38:03.577 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:38:03.577 "hdgst": ${hdgst:-false}, 00:38:03.577 "ddgst": ${ddgst:-false} 00:38:03.577 }, 00:38:03.577 "method": "bdev_nvme_attach_controller" 00:38:03.577 } 00:38:03.577 EOF 00:38:03.577 )") 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # gen_fio_conf 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@54 -- # local file 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1343 -- # local sanitizers 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@56 -- # cat 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # shift 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1347 -- # local asan_lib= 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@582 -- # cat 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file = 1 )) 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # grep libasan 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file <= files )) 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@584 -- # jq . 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@585 -- # IFS=, 00:38:03.577 19:34:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:38:03.577 "params": { 00:38:03.577 "name": "Nvme0", 00:38:03.577 "trtype": "tcp", 00:38:03.577 "traddr": "10.0.0.2", 00:38:03.577 "adrfam": "ipv4", 00:38:03.577 "trsvcid": "4420", 00:38:03.577 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:38:03.577 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:38:03.577 "hdgst": true, 00:38:03.577 "ddgst": true 00:38:03.577 }, 00:38:03.577 "method": "bdev_nvme_attach_controller" 00:38:03.577 }' 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # asan_lib= 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # asan_lib= 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:38:03.577 19:34:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:38:03.839 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:38:03.839 ... 00:38:03.839 fio-3.35 00:38:03.839 Starting 3 threads 00:38:16.078 00:38:16.078 filename0: (groupid=0, jobs=1): err= 0: pid=2666: Tue Nov 26 19:34:27 2024 00:38:16.078 read: IOPS=202, BW=25.3MiB/s (26.5MB/s)(254MiB/10047msec) 00:38:16.078 slat (nsec): min=5872, max=47121, avg=7883.82, stdev=1923.25 00:38:16.078 clat (usec): min=8012, max=56892, avg=14802.56, stdev=3910.16 00:38:16.078 lat (usec): min=8018, max=56900, avg=14810.44, stdev=3910.11 00:38:16.078 clat percentiles (usec): 00:38:16.078 | 1.00th=[10028], 5.00th=[12256], 10.00th=[13042], 20.00th=[13566], 00:38:16.078 | 30.00th=[13829], 40.00th=[14222], 50.00th=[14484], 60.00th=[14746], 00:38:16.078 | 70.00th=[15139], 80.00th=[15533], 90.00th=[16188], 95.00th=[16712], 00:38:16.078 | 99.00th=[17957], 99.50th=[54789], 99.90th=[56361], 99.95th=[56361], 00:38:16.078 | 99.99th=[56886] 00:38:16.078 bw ( KiB/s): min=23552, max=28416, per=31.99%, avg=25984.00, stdev=1236.13, samples=20 00:38:16.078 iops : min= 184, max= 222, avg=203.00, stdev= 9.66, samples=20 00:38:16.078 lat (msec) : 10=1.03%, 20=98.13%, 50=0.05%, 100=0.79% 00:38:16.078 cpu : usr=94.60%, sys=5.15%, ctx=18, majf=0, minf=139 00:38:16.078 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:38:16.078 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:16.078 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:16.078 issued rwts: total=2032,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:38:16.078 latency : target=0, window=0, percentile=100.00%, depth=3 00:38:16.078 filename0: (groupid=0, jobs=1): err= 0: pid=2667: Tue Nov 26 19:34:27 2024 00:38:16.078 read: IOPS=224, BW=28.0MiB/s (29.4MB/s)(282MiB/10045msec) 00:38:16.078 slat (nsec): min=5902, max=44453, avg=9228.21, stdev=2453.53 00:38:16.078 clat (usec): min=7515, max=53029, avg=13350.55, stdev=2233.70 00:38:16.078 lat (usec): min=7529, max=53040, avg=13359.78, stdev=2233.68 00:38:16.078 clat percentiles (usec): 00:38:16.078 | 1.00th=[ 8848], 5.00th=[11076], 10.00th=[11731], 20.00th=[12387], 00:38:16.078 | 30.00th=[12780], 40.00th=[13042], 50.00th=[13304], 60.00th=[13566], 00:38:16.078 | 70.00th=[13960], 80.00th=[14222], 90.00th=[14877], 95.00th=[15270], 00:38:16.078 | 99.00th=[16188], 99.50th=[16909], 99.90th=[52691], 99.95th=[52691], 00:38:16.078 | 99.99th=[53216] 00:38:16.078 bw ( KiB/s): min=26368, max=31744, per=35.46%, avg=28800.00, stdev=1311.94, samples=20 00:38:16.078 iops : min= 206, max= 248, avg=225.00, stdev=10.25, samples=20 00:38:16.078 lat (msec) : 10=2.31%, 20=97.47%, 50=0.04%, 100=0.18% 00:38:16.078 cpu : usr=94.45%, sys=5.27%, ctx=23, majf=0, minf=165 00:38:16.078 IO depths : 1=0.2%, 2=99.8%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:38:16.078 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:16.078 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:16.078 issued rwts: total=2252,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:38:16.078 latency : target=0, window=0, percentile=100.00%, depth=3 00:38:16.078 filename0: (groupid=0, jobs=1): err= 0: pid=2668: Tue Nov 26 19:34:27 2024 00:38:16.078 read: IOPS=208, BW=26.0MiB/s (27.3MB/s)(261MiB/10043msec) 00:38:16.078 slat (nsec): min=5866, max=32270, avg=7381.56, stdev=1646.09 00:38:16.078 clat (usec): min=8429, max=55687, avg=14380.19, stdev=2786.41 00:38:16.078 lat (usec): min=8438, max=55694, avg=14387.57, stdev=2786.39 00:38:16.078 clat percentiles (usec): 00:38:16.078 | 1.00th=[10159], 5.00th=[12125], 10.00th=[12649], 20.00th=[13304], 00:38:16.078 | 30.00th=[13698], 40.00th=[13960], 50.00th=[14222], 60.00th=[14615], 00:38:16.078 | 70.00th=[14877], 80.00th=[15270], 90.00th=[15795], 95.00th=[16319], 00:38:16.078 | 99.00th=[17433], 99.50th=[18744], 99.90th=[55313], 99.95th=[55313], 00:38:16.078 | 99.99th=[55837] 00:38:16.078 bw ( KiB/s): min=24832, max=28160, per=32.92%, avg=26739.20, stdev=820.02, samples=20 00:38:16.078 iops : min= 194, max= 220, avg=208.90, stdev= 6.41, samples=20 00:38:16.078 lat (msec) : 10=0.91%, 20=98.71%, 100=0.38% 00:38:16.078 cpu : usr=95.14%, sys=4.61%, ctx=25, majf=0, minf=130 00:38:16.078 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:38:16.078 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:16.078 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:38:16.078 issued rwts: total=2091,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:38:16.078 latency : target=0, window=0, percentile=100.00%, depth=3 00:38:16.078 00:38:16.078 Run status group 0 (all jobs): 00:38:16.078 READ: bw=79.3MiB/s (83.2MB/s), 25.3MiB/s-28.0MiB/s (26.5MB/s-29.4MB/s), io=797MiB (836MB), run=10043-10047msec 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- target/dif.sh@132 -- # destroy_subsystems 0 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- target/dif.sh@43 -- # local sub 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- target/dif.sh@45 -- # for sub in "$@" 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- target/dif.sh@46 -- # destroy_subsystem 0 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- target/dif.sh@36 -- # local sub_id=0 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:16.078 00:38:16.078 real 0m11.302s 00:38:16.078 user 0m42.945s 00:38:16.078 sys 0m1.865s 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:16.078 19:34:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:38:16.078 ************************************ 00:38:16.078 END TEST fio_dif_digest 00:38:16.078 ************************************ 00:38:16.078 19:34:27 nvmf_dif -- target/dif.sh@146 -- # trap - SIGINT SIGTERM EXIT 00:38:16.078 19:34:27 nvmf_dif -- target/dif.sh@147 -- # nvmftestfini 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@516 -- # nvmfcleanup 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@121 -- # sync 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@124 -- # set +e 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@125 -- # for i in {1..20} 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:38:16.078 rmmod nvme_tcp 00:38:16.078 rmmod nvme_fabrics 00:38:16.078 rmmod nvme_keyring 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@128 -- # set -e 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@129 -- # return 0 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@517 -- # '[' -n 4185591 ']' 00:38:16.078 19:34:27 nvmf_dif -- nvmf/common.sh@518 -- # killprocess 4185591 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@954 -- # '[' -z 4185591 ']' 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@958 -- # kill -0 4185591 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@959 -- # uname 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 4185591 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@972 -- # echo 'killing process with pid 4185591' 00:38:16.078 killing process with pid 4185591 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@973 -- # kill 4185591 00:38:16.078 19:34:27 nvmf_dif -- common/autotest_common.sh@978 -- # wait 4185591 00:38:16.079 19:34:27 nvmf_dif -- nvmf/common.sh@520 -- # '[' iso == iso ']' 00:38:16.079 19:34:27 nvmf_dif -- nvmf/common.sh@521 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:38:19.385 Waiting for block devices as requested 00:38:19.385 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:38:19.385 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:38:19.647 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:38:19.647 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:38:19.647 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:38:19.909 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:38:19.909 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:38:19.909 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:38:20.171 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:38:20.171 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@297 -- # iptr 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@791 -- # iptables-save 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@791 -- # iptables-restore 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@302 -- # remove_spdk_ns 00:38:20.432 19:34:32 nvmf_dif -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:20.432 19:34:32 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:38:20.432 19:34:32 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:22.980 19:34:35 nvmf_dif -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:38:22.980 00:38:22.980 real 1m20.092s 00:38:22.980 user 8m1.360s 00:38:22.980 sys 0m22.734s 00:38:22.980 19:34:35 nvmf_dif -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:22.980 19:34:35 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:38:22.980 ************************************ 00:38:22.980 END TEST nvmf_dif 00:38:22.980 ************************************ 00:38:22.980 19:34:35 -- spdk/autotest.sh@290 -- # run_test nvmf_abort_qd_sizes /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:38:22.980 19:34:35 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:38:22.980 19:34:35 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:22.980 19:34:35 -- common/autotest_common.sh@10 -- # set +x 00:38:22.980 ************************************ 00:38:22.980 START TEST nvmf_abort_qd_sizes 00:38:22.980 ************************************ 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:38:22.980 * Looking for test storage... 00:38:22.980 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1693 -- # lcov --version 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # IFS=.-: 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # read -ra ver1 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # IFS=.-: 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # read -ra ver2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@338 -- # local 'op=<' 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@340 -- # ver1_l=2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@341 -- # ver2_l=1 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@344 -- # case "$op" in 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@345 -- # : 1 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # decimal 1 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=1 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 1 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # ver1[v]=1 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # decimal 2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # ver2[v]=2 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:22.980 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # return 0 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:38:22.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:22.981 --rc genhtml_branch_coverage=1 00:38:22.981 --rc genhtml_function_coverage=1 00:38:22.981 --rc genhtml_legend=1 00:38:22.981 --rc geninfo_all_blocks=1 00:38:22.981 --rc geninfo_unexecuted_blocks=1 00:38:22.981 00:38:22.981 ' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:38:22.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:22.981 --rc genhtml_branch_coverage=1 00:38:22.981 --rc genhtml_function_coverage=1 00:38:22.981 --rc genhtml_legend=1 00:38:22.981 --rc geninfo_all_blocks=1 00:38:22.981 --rc geninfo_unexecuted_blocks=1 00:38:22.981 00:38:22.981 ' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:38:22.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:22.981 --rc genhtml_branch_coverage=1 00:38:22.981 --rc genhtml_function_coverage=1 00:38:22.981 --rc genhtml_legend=1 00:38:22.981 --rc geninfo_all_blocks=1 00:38:22.981 --rc geninfo_unexecuted_blocks=1 00:38:22.981 00:38:22.981 ' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:38:22.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:22.981 --rc genhtml_branch_coverage=1 00:38:22.981 --rc genhtml_function_coverage=1 00:38:22.981 --rc genhtml_legend=1 00:38:22.981 --rc geninfo_all_blocks=1 00:38:22.981 --rc geninfo_unexecuted_blocks=1 00:38:22.981 00:38:22.981 ' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # uname -s 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@15 -- # shopt -s extglob 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- paths/export.sh@5 -- # export PATH 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@51 -- # : 0 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:38:22.981 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@70 -- # nvmftestinit 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@476 -- # prepare_net_devs 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@438 -- # local -g is_hw=no 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # remove_spdk_ns 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- nvmf/common.sh@309 -- # xtrace_disable 00:38:22.981 19:34:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # pci_devs=() 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # local -a pci_devs 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # pci_net_devs=() 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # pci_drivers=() 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # local -A pci_drivers 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # net_devs=() 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # local -ga net_devs 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # e810=() 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # local -ga e810 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # x722=() 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # local -ga x722 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # mlx=() 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # local -ga mlx 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:38:31.119 Found 0000:31:00.0 (0x8086 - 0x159b) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:38:31.119 Found 0000:31:00.1 (0x8086 - 0x159b) 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:31.119 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:38:31.120 Found net devices under 0000:31:00.0: cvl_0_0 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:38:31.120 Found net devices under 0000:31:00.1: cvl_0_1 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # is_hw=yes 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:38:31.120 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:38:31.120 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.670 ms 00:38:31.120 00:38:31.120 --- 10.0.0.2 ping statistics --- 00:38:31.120 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:31.120 rtt min/avg/max/mdev = 0.670/0.670/0.670/0.000 ms 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:38:31.120 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:38:31.120 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.218 ms 00:38:31.120 00:38:31.120 --- 10.0.0.1 ping statistics --- 00:38:31.120 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:31.120 rtt min/avg/max/mdev = 0.218/0.218/0.218/0.000 ms 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@450 -- # return 0 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@478 -- # '[' iso == iso ']' 00:38:31.120 19:34:43 nvmf_abort_qd_sizes -- nvmf/common.sh@479 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:38:35.419 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:38:35.419 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:38:35.419 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:38:35.419 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:38:35.420 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@71 -- # nvmfappstart -m 0xf 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@726 -- # xtrace_disable 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@509 -- # nvmfpid=13136 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@510 -- # waitforlisten 13136 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xf 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@835 -- # '[' -z 13136 ']' 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:35.680 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:35.680 19:34:48 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:38:35.940 [2024-11-26 19:34:48.340483] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:38:35.940 [2024-11-26 19:34:48.340533] [ DPDK EAL parameters: nvmf -c 0xf --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:35.940 [2024-11-26 19:34:48.427162] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:38:35.940 [2024-11-26 19:34:48.464204] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:35.940 [2024-11-26 19:34:48.464236] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:35.940 [2024-11-26 19:34:48.464244] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:35.940 [2024-11-26 19:34:48.464251] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:35.940 [2024-11-26 19:34:48.464256] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:35.940 [2024-11-26 19:34:48.465981] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:35.940 [2024-11-26 19:34:48.466095] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:38:35.940 [2024-11-26 19:34:48.466250] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:38:35.940 [2024-11-26 19:34:48.466250] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@868 -- # return 0 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@732 -- # xtrace_disable 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@73 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini || :; clean_kernel_target' SIGINT SIGTERM EXIT 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # mapfile -t nvmes 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # nvme_in_userspace 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@312 -- # local bdf bdfs 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@313 -- # local nvmes 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@315 -- # [[ -n 0000:65:00.0 ]] 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@316 -- # nvmes=(${pci_bus_cache["0x010802"]}) 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@321 -- # for bdf in "${nvmes[@]}" 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@322 -- # [[ -e /sys/bus/pci/drivers/nvme/0000:65:00.0 ]] 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # uname -s 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # [[ Linux == FreeBSD ]] 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@326 -- # bdfs+=("$bdf") 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@328 -- # (( 1 )) 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- scripts/common.sh@329 -- # printf '%s\n' 0000:65:00.0 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@76 -- # (( 1 > 0 )) 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@78 -- # nvme=0000:65:00.0 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@80 -- # run_test spdk_target_abort spdk_target 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:36.881 19:34:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:38:36.881 ************************************ 00:38:36.881 START TEST spdk_target_abort 00:38:36.881 ************************************ 00:38:36.881 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1129 -- # spdk_target 00:38:36.881 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@43 -- # local name=spdk_target 00:38:36.881 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@45 -- # rpc_cmd bdev_nvme_attach_controller -t pcie -a 0000:65:00.0 -b spdk_target 00:38:36.881 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:36.881 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:37.141 spdk_targetn1 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@47 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:37.141 [2024-11-26 19:34:49.551915] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:testnqn -a -s SPDKISFASTANDAWESOME 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@49 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:testnqn spdk_targetn1 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@50 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:testnqn -t tcp -a 10.0.0.2 -s 4420 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:37.141 [2024-11-26 19:34:49.608245] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@52 -- # rabort tcp IPv4 10.0.0.2 4420 nqn.2016-06.io.spdk:testnqn 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.2 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:38:37.141 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2' 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:38:37.142 19:34:49 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:38:37.402 [2024-11-26 19:34:49.858288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:776 len:8 PRP1 0x200004ac4000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.858317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:0062 p:1 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.875839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:1448 len:8 PRP1 0x200004ac4000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.875857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:00b6 p:1 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.882339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:189 nsid:1 lba:1616 len:8 PRP1 0x200004ac0000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.882356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:189 cdw0:0 sqhd:00cc p:1 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.897269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:2128 len:8 PRP1 0x200004abe000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.897286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.897592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:2152 len:8 PRP1 0x200004ac6000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.897605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.900650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:2352 len:8 PRP1 0x200004abe000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.900666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.905304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:2392 len:8 PRP1 0x200004ac0000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.905321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.905418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:189 nsid:1 lba:2408 len:8 PRP1 0x200004ac4000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.905428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:189 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.919018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:2896 len:8 PRP1 0x200004abe000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.919035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.927337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:3192 len:8 PRP1 0x200004ac6000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.927359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:0091 p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.935294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:3440 len:8 PRP1 0x200004ac6000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.935310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:00af p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.951371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:3960 len:8 PRP1 0x200004ac6000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.951390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:00f0 p:0 m:0 dnr:0 00:38:37.402 [2024-11-26 19:34:49.959212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:4200 len:8 PRP1 0x200004ac0000 PRP2 0x0 00:38:37.402 [2024-11-26 19:34:49.959231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:38:40.701 Initializing NVMe Controllers 00:38:40.701 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:38:40.701 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:38:40.701 Initialization complete. Launching workers. 00:38:40.701 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 11428, failed: 13 00:38:40.701 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 3169, failed to submit 8272 00:38:40.701 success 703, unsuccessful 2466, failed 0 00:38:40.701 19:34:52 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:38:40.701 19:34:52 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:38:40.701 [2024-11-26 19:34:53.057198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:181 nsid:1 lba:488 len:8 PRP1 0x200004e50000 PRP2 0x0 00:38:40.701 [2024-11-26 19:34:53.057244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:181 cdw0:0 sqhd:0048 p:1 m:0 dnr:0 00:38:40.701 [2024-11-26 19:34:53.081032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:178 nsid:1 lba:824 len:8 PRP1 0x200004e42000 PRP2 0x0 00:38:40.701 [2024-11-26 19:34:53.081057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:178 cdw0:0 sqhd:0076 p:1 m:0 dnr:0 00:38:40.701 [2024-11-26 19:34:53.097009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:168 nsid:1 lba:1192 len:8 PRP1 0x200004e48000 PRP2 0x0 00:38:40.701 [2024-11-26 19:34:53.097033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:168 cdw0:0 sqhd:009f p:1 m:0 dnr:0 00:38:40.701 [2024-11-26 19:34:53.151975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:180 nsid:1 lba:2528 len:8 PRP1 0x200004e50000 PRP2 0x0 00:38:40.701 [2024-11-26 19:34:53.152000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:180 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:38:40.701 [2024-11-26 19:34:53.221008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:188 nsid:1 lba:4216 len:8 PRP1 0x200004e58000 PRP2 0x0 00:38:40.701 [2024-11-26 19:34:53.221031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:188 cdw0:0 sqhd:001a p:1 m:0 dnr:0 00:38:40.701 [2024-11-26 19:34:53.243640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:176 nsid:1 lba:4816 len:8 PRP1 0x200004e3a000 PRP2 0x0 00:38:40.701 [2024-11-26 19:34:53.243663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:176 cdw0:0 sqhd:005b p:1 m:0 dnr:0 00:38:43.999 Initializing NVMe Controllers 00:38:43.999 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:38:43.999 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:38:43.999 Initialization complete. Launching workers. 00:38:43.999 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 8559, failed: 6 00:38:43.999 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1191, failed to submit 7374 00:38:43.999 success 361, unsuccessful 830, failed 0 00:38:43.999 19:34:56 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:38:43.999 19:34:56 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:38:47.301 Initializing NVMe Controllers 00:38:47.302 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:38:47.302 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:38:47.302 Initialization complete. Launching workers. 00:38:47.302 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 41805, failed: 0 00:38:47.302 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2671, failed to submit 39134 00:38:47.302 success 597, unsuccessful 2074, failed 0 00:38:47.302 19:34:59 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@54 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:testnqn 00:38:47.302 19:34:59 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.302 19:34:59 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:47.302 19:34:59 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:47.302 19:34:59 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@55 -- # rpc_cmd bdev_nvme_detach_controller spdk_target 00:38:47.302 19:34:59 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.302 19:34:59 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:48.687 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:48.687 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@61 -- # killprocess 13136 00:38:48.687 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@954 -- # '[' -z 13136 ']' 00:38:48.687 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@958 -- # kill -0 13136 00:38:48.687 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@959 -- # uname 00:38:48.687 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:48.687 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 13136 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@972 -- # echo 'killing process with pid 13136' 00:38:48.948 killing process with pid 13136 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@973 -- # kill 13136 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@978 -- # wait 13136 00:38:48.948 00:38:48.948 real 0m12.209s 00:38:48.948 user 0m49.891s 00:38:48.948 sys 0m1.827s 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:38:48.948 ************************************ 00:38:48.948 END TEST spdk_target_abort 00:38:48.948 ************************************ 00:38:48.948 19:35:01 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@81 -- # run_test kernel_target_abort kernel_target 00:38:48.948 19:35:01 nvmf_abort_qd_sizes -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:38:48.948 19:35:01 nvmf_abort_qd_sizes -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:48.948 19:35:01 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:38:48.948 ************************************ 00:38:48.948 START TEST kernel_target_abort 00:38:48.948 ************************************ 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1129 -- # kernel_target 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # get_main_ns_ip 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@769 -- # local ip 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@770 -- # ip_candidates=() 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@770 -- # local -A ip_candidates 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@660 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@662 -- # nvmet=/sys/kernel/config/nvmet 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@663 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@664 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@665 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@667 -- # local block nvme 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@669 -- # [[ ! -e /sys/module/nvmet ]] 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@670 -- # modprobe nvmet 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet ]] 00:38:48.948 19:35:01 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@675 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:38:53.155 Waiting for block devices as requested 00:38:53.155 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:38:53.155 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:38:53.155 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:38:53.155 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:38:53.156 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:38:53.156 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:38:53.156 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:38:53.156 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:38:53.416 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:38:53.416 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:38:53.677 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:38:53.677 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:38:53.677 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:38:53.677 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:38:53.938 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:38:53.938 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:38:53.938 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@678 -- # for block in /sys/block/nvme* 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@679 -- # [[ -e /sys/block/nvme0n1 ]] 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@680 -- # is_block_zoned nvme0n1 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@681 -- # block_in_use nvme0n1 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:38:54.199 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:38:54.459 No valid GPT data, bailing 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # pt= 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@395 -- # return 1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@681 -- # nvme=/dev/nvme0n1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@684 -- # [[ -b /dev/nvme0n1 ]] 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@687 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@688 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@693 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@695 -- # echo 1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@696 -- # echo /dev/nvme0n1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@697 -- # echo 1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@699 -- # echo 10.0.0.1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@700 -- # echo tcp 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@701 -- # echo 4420 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@702 -- # echo ipv4 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@705 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@708 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 --hostid=00539ede-7deb-ec11-9bc7-a4bf01928396 -a 10.0.0.1 -t tcp -s 4420 00:38:54.459 00:38:54.459 Discovery Log Number of Records 2, Generation counter 2 00:38:54.459 =====Discovery Log Entry 0====== 00:38:54.459 trtype: tcp 00:38:54.459 adrfam: ipv4 00:38:54.459 subtype: current discovery subsystem 00:38:54.459 treq: not specified, sq flow control disable supported 00:38:54.459 portid: 1 00:38:54.459 trsvcid: 4420 00:38:54.459 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:38:54.459 traddr: 10.0.0.1 00:38:54.459 eflags: none 00:38:54.459 sectype: none 00:38:54.459 =====Discovery Log Entry 1====== 00:38:54.459 trtype: tcp 00:38:54.459 adrfam: ipv4 00:38:54.459 subtype: nvme subsystem 00:38:54.459 treq: not specified, sq flow control disable supported 00:38:54.459 portid: 1 00:38:54.459 trsvcid: 4420 00:38:54.459 subnqn: nqn.2016-06.io.spdk:testnqn 00:38:54.459 traddr: 10.0.0.1 00:38:54.459 eflags: none 00:38:54.459 sectype: none 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@66 -- # rabort tcp IPv4 10.0.0.1 4420 nqn.2016-06.io.spdk:testnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.1 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1' 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420' 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:38:54.459 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:38:54.460 19:35:06 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:38:57.759 Initializing NVMe Controllers 00:38:57.759 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:38:57.759 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:38:57.759 Initialization complete. Launching workers. 00:38:57.759 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 66799, failed: 0 00:38:57.759 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 66799, failed to submit 0 00:38:57.759 success 0, unsuccessful 66799, failed 0 00:38:57.759 19:35:10 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:38:57.759 19:35:10 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:39:01.057 Initializing NVMe Controllers 00:39:01.057 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:39:01.057 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:39:01.057 Initialization complete. Launching workers. 00:39:01.057 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 107624, failed: 0 00:39:01.057 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 27062, failed to submit 80562 00:39:01.057 success 0, unsuccessful 27062, failed 0 00:39:01.057 19:35:13 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:39:01.057 19:35:13 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:39:04.384 Initializing NVMe Controllers 00:39:04.384 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:39:04.384 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:39:04.384 Initialization complete. Launching workers. 00:39:04.384 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 101581, failed: 0 00:39:04.384 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 25410, failed to submit 76171 00:39:04.384 success 0, unsuccessful 25410, failed 0 00:39:04.384 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@67 -- # clean_kernel_target 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@712 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@714 -- # echo 0 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@716 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@718 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@719 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@721 -- # modules=(/sys/module/nvmet/holders/*) 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@723 -- # modprobe -r nvmet_tcp nvmet 00:39:04.385 19:35:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@726 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:39:06.928 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:39:06.928 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:39:06.928 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:39:06.928 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:39:07.189 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:39:09.099 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:39:09.359 00:39:09.359 real 0m20.284s 00:39:09.359 user 0m9.647s 00:39:09.359 sys 0m6.066s 00:39:09.359 19:35:21 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:39:09.359 19:35:21 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@10 -- # set +x 00:39:09.359 ************************************ 00:39:09.359 END TEST kernel_target_abort 00:39:09.359 ************************************ 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@84 -- # nvmftestfini 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@516 -- # nvmfcleanup 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@121 -- # sync 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@124 -- # set +e 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@125 -- # for i in {1..20} 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:39:09.359 rmmod nvme_tcp 00:39:09.359 rmmod nvme_fabrics 00:39:09.359 rmmod nvme_keyring 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@128 -- # set -e 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@129 -- # return 0 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@517 -- # '[' -n 13136 ']' 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@518 -- # killprocess 13136 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- common/autotest_common.sh@954 -- # '[' -z 13136 ']' 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- common/autotest_common.sh@958 -- # kill -0 13136 00:39:09.359 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (13136) - No such process 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- common/autotest_common.sh@981 -- # echo 'Process with pid 13136 is not found' 00:39:09.359 Process with pid 13136 is not found 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@520 -- # '[' iso == iso ']' 00:39:09.359 19:35:21 nvmf_abort_qd_sizes -- nvmf/common.sh@521 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:39:13.567 Waiting for block devices as requested 00:39:13.567 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:39:13.567 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:39:13.827 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:39:13.827 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:39:14.088 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:39:14.088 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:39:14.088 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:39:14.088 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:39:14.348 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:39:14.348 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@297 -- # iptr 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@791 -- # iptables-save 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@791 -- # iptables-restore 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@302 -- # remove_spdk_ns 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:39:14.609 19:35:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:39:17.153 19:35:29 nvmf_abort_qd_sizes -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:39:17.153 00:39:17.153 real 0m54.136s 00:39:17.153 user 1m5.528s 00:39:17.153 sys 0m20.178s 00:39:17.153 19:35:29 nvmf_abort_qd_sizes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:39:17.153 19:35:29 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:39:17.153 ************************************ 00:39:17.153 END TEST nvmf_abort_qd_sizes 00:39:17.153 ************************************ 00:39:17.153 19:35:29 -- spdk/autotest.sh@292 -- # run_test keyring_file /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:39:17.153 19:35:29 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:39:17.153 19:35:29 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:39:17.153 19:35:29 -- common/autotest_common.sh@10 -- # set +x 00:39:17.153 ************************************ 00:39:17.153 START TEST keyring_file 00:39:17.153 ************************************ 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:39:17.153 * Looking for test storage... 00:39:17.153 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1693 -- # lcov --version 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@333 -- # local ver1 ver1_l 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@334 -- # local ver2 ver2_l 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@336 -- # IFS=.-: 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@336 -- # read -ra ver1 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@337 -- # IFS=.-: 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@337 -- # read -ra ver2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@338 -- # local 'op=<' 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@340 -- # ver1_l=2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@341 -- # ver2_l=1 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@344 -- # case "$op" in 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@345 -- # : 1 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@364 -- # (( v = 0 )) 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@365 -- # decimal 1 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@353 -- # local d=1 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@355 -- # echo 1 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@365 -- # ver1[v]=1 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@366 -- # decimal 2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@353 -- # local d=2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@355 -- # echo 2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@366 -- # ver2[v]=2 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:39:17.153 19:35:29 keyring_file -- scripts/common.sh@368 -- # return 0 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:39:17.153 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:17.153 --rc genhtml_branch_coverage=1 00:39:17.153 --rc genhtml_function_coverage=1 00:39:17.153 --rc genhtml_legend=1 00:39:17.153 --rc geninfo_all_blocks=1 00:39:17.153 --rc geninfo_unexecuted_blocks=1 00:39:17.153 00:39:17.153 ' 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:39:17.153 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:17.153 --rc genhtml_branch_coverage=1 00:39:17.153 --rc genhtml_function_coverage=1 00:39:17.153 --rc genhtml_legend=1 00:39:17.153 --rc geninfo_all_blocks=1 00:39:17.153 --rc geninfo_unexecuted_blocks=1 00:39:17.153 00:39:17.153 ' 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:39:17.153 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:17.153 --rc genhtml_branch_coverage=1 00:39:17.153 --rc genhtml_function_coverage=1 00:39:17.153 --rc genhtml_legend=1 00:39:17.153 --rc geninfo_all_blocks=1 00:39:17.153 --rc geninfo_unexecuted_blocks=1 00:39:17.153 00:39:17.153 ' 00:39:17.153 19:35:29 keyring_file -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:39:17.153 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:17.153 --rc genhtml_branch_coverage=1 00:39:17.153 --rc genhtml_function_coverage=1 00:39:17.153 --rc genhtml_legend=1 00:39:17.153 --rc geninfo_all_blocks=1 00:39:17.153 --rc geninfo_unexecuted_blocks=1 00:39:17.153 00:39:17.153 ' 00:39:17.153 19:35:29 keyring_file -- keyring/file.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:39:17.153 19:35:29 keyring_file -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:39:17.153 19:35:29 keyring_file -- nvmf/common.sh@7 -- # uname -s 00:39:17.153 19:35:29 keyring_file -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:39:17.153 19:35:29 keyring_file -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:39:17.153 19:35:29 keyring_file -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:39:17.153 19:35:29 keyring_file -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:39:17.154 19:35:29 keyring_file -- scripts/common.sh@15 -- # shopt -s extglob 00:39:17.154 19:35:29 keyring_file -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:39:17.154 19:35:29 keyring_file -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:39:17.154 19:35:29 keyring_file -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:39:17.154 19:35:29 keyring_file -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:17.154 19:35:29 keyring_file -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:17.154 19:35:29 keyring_file -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:17.154 19:35:29 keyring_file -- paths/export.sh@5 -- # export PATH 00:39:17.154 19:35:29 keyring_file -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@51 -- # : 0 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:39:17.154 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@55 -- # have_pci_nics=0 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@13 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@14 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@15 -- # key0=00112233445566778899aabbccddeeff 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@16 -- # key1=112233445566778899aabbccddeeff00 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@24 -- # trap cleanup EXIT 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@26 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@17 -- # name=key0 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@17 -- # digest=0 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@18 -- # mktemp 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.IHFmCEOG6R 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@730 -- # local prefix key digest 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@732 -- # digest=0 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@733 -- # python - 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.IHFmCEOG6R 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.IHFmCEOG6R 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@26 -- # key0path=/tmp/tmp.IHFmCEOG6R 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@27 -- # prep_key key1 112233445566778899aabbccddeeff00 0 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@17 -- # name=key1 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@17 -- # digest=0 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@18 -- # mktemp 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.tw5Rv6bC7R 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@730 -- # local prefix key digest 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@732 -- # key=112233445566778899aabbccddeeff00 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@732 -- # digest=0 00:39:17.154 19:35:29 keyring_file -- nvmf/common.sh@733 -- # python - 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.tw5Rv6bC7R 00:39:17.154 19:35:29 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.tw5Rv6bC7R 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@27 -- # key1path=/tmp/tmp.tw5Rv6bC7R 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@30 -- # tgtpid=24021 00:39:17.154 19:35:29 keyring_file -- keyring/file.sh@32 -- # waitforlisten 24021 00:39:17.154 19:35:29 keyring_file -- common/autotest_common.sh@835 -- # '[' -z 24021 ']' 00:39:17.154 19:35:29 keyring_file -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:39:17.154 19:35:29 keyring_file -- common/autotest_common.sh@840 -- # local max_retries=100 00:39:17.154 19:35:29 keyring_file -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:39:17.154 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:39:17.154 19:35:29 keyring_file -- common/autotest_common.sh@844 -- # xtrace_disable 00:39:17.154 19:35:29 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:39:17.154 [2024-11-26 19:35:29.706837] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:39:17.154 [2024-11-26 19:35:29.706921] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid24021 ] 00:39:17.413 [2024-11-26 19:35:29.789993] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:39:17.413 [2024-11-26 19:35:29.831736] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@868 -- # return 0 00:39:17.984 19:35:30 keyring_file -- keyring/file.sh@33 -- # rpc_cmd 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:39:17.984 [2024-11-26 19:35:30.494424] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:39:17.984 null0 00:39:17.984 [2024-11-26 19:35:30.526474] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:39:17.984 [2024-11-26 19:35:30.526779] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:17.984 19:35:30 keyring_file -- keyring/file.sh@44 -- # NOT rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@655 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:39:17.984 [2024-11-26 19:35:30.554537] nvmf_rpc.c: 762:nvmf_rpc_listen_paused: *ERROR*: Listener already exists 00:39:17.984 request: 00:39:17.984 { 00:39:17.984 "nqn": "nqn.2016-06.io.spdk:cnode0", 00:39:17.984 "secure_channel": false, 00:39:17.984 "listen_address": { 00:39:17.984 "trtype": "tcp", 00:39:17.984 "traddr": "127.0.0.1", 00:39:17.984 "trsvcid": "4420" 00:39:17.984 }, 00:39:17.984 "method": "nvmf_subsystem_add_listener", 00:39:17.984 "req_id": 1 00:39:17.984 } 00:39:17.984 Got JSON-RPC error response 00:39:17.984 response: 00:39:17.984 { 00:39:17.984 "code": -32602, 00:39:17.984 "message": "Invalid parameters" 00:39:17.984 } 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:39:17.984 19:35:30 keyring_file -- keyring/file.sh@47 -- # bperfpid=24042 00:39:17.984 19:35:30 keyring_file -- keyring/file.sh@49 -- # waitforlisten 24042 /var/tmp/bperf.sock 00:39:17.984 19:35:30 keyring_file -- keyring/file.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@835 -- # '[' -z 24042 ']' 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@840 -- # local max_retries=100 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:39:17.984 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@844 -- # xtrace_disable 00:39:17.984 19:35:30 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:39:18.244 [2024-11-26 19:35:30.611120] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:39:18.244 [2024-11-26 19:35:30.611170] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid24042 ] 00:39:18.244 [2024-11-26 19:35:30.706090] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:39:18.244 [2024-11-26 19:35:30.742437] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:39:18.813 19:35:31 keyring_file -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:39:18.813 19:35:31 keyring_file -- common/autotest_common.sh@868 -- # return 0 00:39:18.813 19:35:31 keyring_file -- keyring/file.sh@50 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:18.814 19:35:31 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:19.073 19:35:31 keyring_file -- keyring/file.sh@51 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.tw5Rv6bC7R 00:39:19.073 19:35:31 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.tw5Rv6bC7R 00:39:19.333 19:35:31 keyring_file -- keyring/file.sh@52 -- # get_key key0 00:39:19.333 19:35:31 keyring_file -- keyring/file.sh@52 -- # jq -r .path 00:39:19.333 19:35:31 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:19.333 19:35:31 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:19.333 19:35:31 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:19.333 19:35:31 keyring_file -- keyring/file.sh@52 -- # [[ /tmp/tmp.IHFmCEOG6R == \/\t\m\p\/\t\m\p\.\I\H\F\m\C\E\O\G\6\R ]] 00:39:19.333 19:35:31 keyring_file -- keyring/file.sh@53 -- # get_key key1 00:39:19.333 19:35:31 keyring_file -- keyring/file.sh@53 -- # jq -r .path 00:39:19.333 19:35:31 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:19.333 19:35:31 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:39:19.333 19:35:31 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:19.594 19:35:32 keyring_file -- keyring/file.sh@53 -- # [[ /tmp/tmp.tw5Rv6bC7R == \/\t\m\p\/\t\m\p\.\t\w\5\R\v\6\b\C\7\R ]] 00:39:19.594 19:35:32 keyring_file -- keyring/file.sh@54 -- # get_refcnt key0 00:39:19.594 19:35:32 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:19.594 19:35:32 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:19.594 19:35:32 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:19.594 19:35:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:19.594 19:35:32 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:19.854 19:35:32 keyring_file -- keyring/file.sh@54 -- # (( 1 == 1 )) 00:39:19.854 19:35:32 keyring_file -- keyring/file.sh@55 -- # get_refcnt key1 00:39:19.854 19:35:32 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:39:19.854 19:35:32 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:19.854 19:35:32 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:19.854 19:35:32 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:39:19.854 19:35:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:19.854 19:35:32 keyring_file -- keyring/file.sh@55 -- # (( 1 == 1 )) 00:39:19.854 19:35:32 keyring_file -- keyring/file.sh@58 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:19.854 19:35:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:20.114 [2024-11-26 19:35:32.586166] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:39:20.114 nvme0n1 00:39:20.114 19:35:32 keyring_file -- keyring/file.sh@60 -- # get_refcnt key0 00:39:20.114 19:35:32 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:20.114 19:35:32 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:20.114 19:35:32 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:20.114 19:35:32 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:20.114 19:35:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:20.373 19:35:32 keyring_file -- keyring/file.sh@60 -- # (( 2 == 2 )) 00:39:20.373 19:35:32 keyring_file -- keyring/file.sh@61 -- # get_refcnt key1 00:39:20.373 19:35:32 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:39:20.373 19:35:32 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:20.373 19:35:32 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:20.373 19:35:32 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:39:20.373 19:35:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:20.633 19:35:33 keyring_file -- keyring/file.sh@61 -- # (( 1 == 1 )) 00:39:20.633 19:35:33 keyring_file -- keyring/file.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:39:20.633 Running I/O for 1 seconds... 00:39:21.573 14728.00 IOPS, 57.53 MiB/s 00:39:21.573 Latency(us) 00:39:21.573 [2024-11-26T18:35:34.198Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:39:21.573 Job: nvme0n1 (Core Mask 0x2, workload: randrw, percentage: 50, depth: 128, IO size: 4096) 00:39:21.573 nvme0n1 : 1.01 14775.14 57.72 0.00 0.00 8644.02 3659.09 12943.36 00:39:21.573 [2024-11-26T18:35:34.198Z] =================================================================================================================== 00:39:21.573 [2024-11-26T18:35:34.198Z] Total : 14775.14 57.72 0.00 0.00 8644.02 3659.09 12943.36 00:39:21.573 { 00:39:21.573 "results": [ 00:39:21.573 { 00:39:21.573 "job": "nvme0n1", 00:39:21.573 "core_mask": "0x2", 00:39:21.573 "workload": "randrw", 00:39:21.573 "percentage": 50, 00:39:21.573 "status": "finished", 00:39:21.573 "queue_depth": 128, 00:39:21.573 "io_size": 4096, 00:39:21.573 "runtime": 1.005608, 00:39:21.573 "iops": 14775.141009220293, 00:39:21.573 "mibps": 57.71539456726677, 00:39:21.573 "io_failed": 0, 00:39:21.573 "io_timeout": 0, 00:39:21.573 "avg_latency_us": 8644.022922780096, 00:39:21.573 "min_latency_us": 3659.0933333333332, 00:39:21.573 "max_latency_us": 12943.36 00:39:21.573 } 00:39:21.573 ], 00:39:21.573 "core_count": 1 00:39:21.573 } 00:39:21.573 19:35:34 keyring_file -- keyring/file.sh@65 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:39:21.574 19:35:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:39:21.833 19:35:34 keyring_file -- keyring/file.sh@66 -- # get_refcnt key0 00:39:21.833 19:35:34 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:21.833 19:35:34 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:21.833 19:35:34 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:21.833 19:35:34 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:21.833 19:35:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:22.092 19:35:34 keyring_file -- keyring/file.sh@66 -- # (( 1 == 1 )) 00:39:22.092 19:35:34 keyring_file -- keyring/file.sh@67 -- # get_refcnt key1 00:39:22.092 19:35:34 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:39:22.092 19:35:34 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:22.092 19:35:34 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:22.093 19:35:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:22.093 19:35:34 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:39:22.093 19:35:34 keyring_file -- keyring/file.sh@67 -- # (( 1 == 1 )) 00:39:22.093 19:35:34 keyring_file -- keyring/file.sh@70 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:39:22.093 19:35:34 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:39:22.093 19:35:34 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:39:22.093 19:35:34 keyring_file -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:39:22.093 19:35:34 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:22.093 19:35:34 keyring_file -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:39:22.093 19:35:34 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:22.093 19:35:34 keyring_file -- common/autotest_common.sh@655 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:39:22.093 19:35:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:39:22.353 [2024-11-26 19:35:34.819770] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:39:22.353 [2024-11-26 19:35:34.820687] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14e99f0 (107): Transport endpoint is not connected 00:39:22.353 [2024-11-26 19:35:34.821680] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14e99f0 (9): Bad file descriptor 00:39:22.353 [2024-11-26 19:35:34.822682] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] Ctrlr is in error state 00:39:22.353 [2024-11-26 19:35:34.822692] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:39:22.353 [2024-11-26 19:35:34.822697] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:39:22.353 [2024-11-26 19:35:34.822704] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] in failed state. 00:39:22.353 request: 00:39:22.353 { 00:39:22.353 "name": "nvme0", 00:39:22.353 "trtype": "tcp", 00:39:22.353 "traddr": "127.0.0.1", 00:39:22.353 "adrfam": "ipv4", 00:39:22.353 "trsvcid": "4420", 00:39:22.353 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:22.353 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:22.353 "prchk_reftag": false, 00:39:22.353 "prchk_guard": false, 00:39:22.353 "hdgst": false, 00:39:22.353 "ddgst": false, 00:39:22.353 "psk": "key1", 00:39:22.353 "allow_unrecognized_csi": false, 00:39:22.353 "method": "bdev_nvme_attach_controller", 00:39:22.353 "req_id": 1 00:39:22.353 } 00:39:22.353 Got JSON-RPC error response 00:39:22.353 response: 00:39:22.353 { 00:39:22.353 "code": -5, 00:39:22.353 "message": "Input/output error" 00:39:22.353 } 00:39:22.353 19:35:34 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:39:22.353 19:35:34 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:39:22.353 19:35:34 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:39:22.353 19:35:34 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:39:22.353 19:35:34 keyring_file -- keyring/file.sh@72 -- # get_refcnt key0 00:39:22.353 19:35:34 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:22.353 19:35:34 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:22.353 19:35:34 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:22.353 19:35:34 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:22.353 19:35:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:22.614 19:35:35 keyring_file -- keyring/file.sh@72 -- # (( 1 == 1 )) 00:39:22.614 19:35:35 keyring_file -- keyring/file.sh@73 -- # get_refcnt key1 00:39:22.614 19:35:35 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:39:22.614 19:35:35 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:22.614 19:35:35 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:22.614 19:35:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:22.614 19:35:35 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:39:22.614 19:35:35 keyring_file -- keyring/file.sh@73 -- # (( 1 == 1 )) 00:39:22.614 19:35:35 keyring_file -- keyring/file.sh@76 -- # bperf_cmd keyring_file_remove_key key0 00:39:22.614 19:35:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:39:22.875 19:35:35 keyring_file -- keyring/file.sh@77 -- # bperf_cmd keyring_file_remove_key key1 00:39:22.875 19:35:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key1 00:39:23.134 19:35:35 keyring_file -- keyring/file.sh@78 -- # bperf_cmd keyring_get_keys 00:39:23.134 19:35:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:23.134 19:35:35 keyring_file -- keyring/file.sh@78 -- # jq length 00:39:23.134 19:35:35 keyring_file -- keyring/file.sh@78 -- # (( 0 == 0 )) 00:39:23.134 19:35:35 keyring_file -- keyring/file.sh@81 -- # chmod 0660 /tmp/tmp.IHFmCEOG6R 00:39:23.134 19:35:35 keyring_file -- keyring/file.sh@82 -- # NOT bperf_cmd keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:23.134 19:35:35 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:39:23.134 19:35:35 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:23.134 19:35:35 keyring_file -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:39:23.135 19:35:35 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:23.135 19:35:35 keyring_file -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:39:23.135 19:35:35 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:23.135 19:35:35 keyring_file -- common/autotest_common.sh@655 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:23.135 19:35:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:23.394 [2024-11-26 19:35:35.852357] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.IHFmCEOG6R': 0100660 00:39:23.394 [2024-11-26 19:35:35.852377] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:39:23.394 request: 00:39:23.394 { 00:39:23.394 "name": "key0", 00:39:23.394 "path": "/tmp/tmp.IHFmCEOG6R", 00:39:23.394 "method": "keyring_file_add_key", 00:39:23.394 "req_id": 1 00:39:23.394 } 00:39:23.394 Got JSON-RPC error response 00:39:23.394 response: 00:39:23.394 { 00:39:23.394 "code": -1, 00:39:23.394 "message": "Operation not permitted" 00:39:23.394 } 00:39:23.394 19:35:35 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:39:23.394 19:35:35 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:39:23.394 19:35:35 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:39:23.394 19:35:35 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:39:23.394 19:35:35 keyring_file -- keyring/file.sh@85 -- # chmod 0600 /tmp/tmp.IHFmCEOG6R 00:39:23.394 19:35:35 keyring_file -- keyring/file.sh@86 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:23.394 19:35:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.IHFmCEOG6R 00:39:23.659 19:35:36 keyring_file -- keyring/file.sh@87 -- # rm -f /tmp/tmp.IHFmCEOG6R 00:39:23.659 19:35:36 keyring_file -- keyring/file.sh@89 -- # get_refcnt key0 00:39:23.659 19:35:36 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:23.659 19:35:36 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:23.659 19:35:36 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:23.659 19:35:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:23.659 19:35:36 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:23.659 19:35:36 keyring_file -- keyring/file.sh@89 -- # (( 1 == 1 )) 00:39:23.659 19:35:36 keyring_file -- keyring/file.sh@91 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:23.659 19:35:36 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:39:23.659 19:35:36 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:23.659 19:35:36 keyring_file -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:39:23.659 19:35:36 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:23.659 19:35:36 keyring_file -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:39:23.659 19:35:36 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:23.659 19:35:36 keyring_file -- common/autotest_common.sh@655 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:23.659 19:35:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:23.920 [2024-11-26 19:35:36.361652] keyring.c: 31:keyring_file_check_path: *ERROR*: Could not stat key file '/tmp/tmp.IHFmCEOG6R': No such file or directory 00:39:23.920 [2024-11-26 19:35:36.361669] nvme_tcp.c:2498:nvme_tcp_generate_tls_credentials: *ERROR*: Failed to obtain key 'key0': No such file or directory 00:39:23.920 [2024-11-26 19:35:36.361682] nvme.c: 682:nvme_ctrlr_probe: *ERROR*: Failed to construct NVMe controller for SSD: 127.0.0.1 00:39:23.920 [2024-11-26 19:35:36.361688] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, No such device 00:39:23.920 [2024-11-26 19:35:36.361694] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:39:23.920 [2024-11-26 19:35:36.361699] bdev_nvme.c:6769:spdk_bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 127.0.0.1) 00:39:23.920 request: 00:39:23.920 { 00:39:23.920 "name": "nvme0", 00:39:23.920 "trtype": "tcp", 00:39:23.920 "traddr": "127.0.0.1", 00:39:23.920 "adrfam": "ipv4", 00:39:23.920 "trsvcid": "4420", 00:39:23.920 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:23.920 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:23.920 "prchk_reftag": false, 00:39:23.920 "prchk_guard": false, 00:39:23.920 "hdgst": false, 00:39:23.920 "ddgst": false, 00:39:23.920 "psk": "key0", 00:39:23.920 "allow_unrecognized_csi": false, 00:39:23.920 "method": "bdev_nvme_attach_controller", 00:39:23.920 "req_id": 1 00:39:23.920 } 00:39:23.920 Got JSON-RPC error response 00:39:23.920 response: 00:39:23.920 { 00:39:23.920 "code": -19, 00:39:23.920 "message": "No such device" 00:39:23.920 } 00:39:23.920 19:35:36 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:39:23.920 19:35:36 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:39:23.920 19:35:36 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:39:23.920 19:35:36 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:39:23.920 19:35:36 keyring_file -- keyring/file.sh@93 -- # bperf_cmd keyring_file_remove_key key0 00:39:23.920 19:35:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:39:23.920 19:35:36 keyring_file -- keyring/file.sh@96 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:39:23.920 19:35:36 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@17 -- # name=key0 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@17 -- # digest=0 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@18 -- # mktemp 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.BkUmphVT3Y 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:39:24.181 19:35:36 keyring_file -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:39:24.181 19:35:36 keyring_file -- nvmf/common.sh@730 -- # local prefix key digest 00:39:24.181 19:35:36 keyring_file -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:39:24.181 19:35:36 keyring_file -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:39:24.181 19:35:36 keyring_file -- nvmf/common.sh@732 -- # digest=0 00:39:24.181 19:35:36 keyring_file -- nvmf/common.sh@733 -- # python - 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.BkUmphVT3Y 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.BkUmphVT3Y 00:39:24.181 19:35:36 keyring_file -- keyring/file.sh@96 -- # key0path=/tmp/tmp.BkUmphVT3Y 00:39:24.181 19:35:36 keyring_file -- keyring/file.sh@97 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.BkUmphVT3Y 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.BkUmphVT3Y 00:39:24.181 19:35:36 keyring_file -- keyring/file.sh@98 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:24.181 19:35:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:24.440 nvme0n1 00:39:24.440 19:35:37 keyring_file -- keyring/file.sh@100 -- # get_refcnt key0 00:39:24.440 19:35:37 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:24.440 19:35:37 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:24.440 19:35:37 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:24.440 19:35:37 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:24.440 19:35:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:24.699 19:35:37 keyring_file -- keyring/file.sh@100 -- # (( 2 == 2 )) 00:39:24.699 19:35:37 keyring_file -- keyring/file.sh@101 -- # bperf_cmd keyring_file_remove_key key0 00:39:24.699 19:35:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:39:24.959 19:35:37 keyring_file -- keyring/file.sh@102 -- # get_key key0 00:39:24.959 19:35:37 keyring_file -- keyring/file.sh@102 -- # jq -r .removed 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:24.959 19:35:37 keyring_file -- keyring/file.sh@102 -- # [[ true == \t\r\u\e ]] 00:39:24.959 19:35:37 keyring_file -- keyring/file.sh@103 -- # get_refcnt key0 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:24.959 19:35:37 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:25.219 19:35:37 keyring_file -- keyring/file.sh@103 -- # (( 1 == 1 )) 00:39:25.219 19:35:37 keyring_file -- keyring/file.sh@104 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:39:25.219 19:35:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:39:25.479 19:35:37 keyring_file -- keyring/file.sh@105 -- # bperf_cmd keyring_get_keys 00:39:25.479 19:35:37 keyring_file -- keyring/file.sh@105 -- # jq length 00:39:25.479 19:35:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:25.479 19:35:38 keyring_file -- keyring/file.sh@105 -- # (( 0 == 0 )) 00:39:25.480 19:35:38 keyring_file -- keyring/file.sh@108 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.BkUmphVT3Y 00:39:25.480 19:35:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.BkUmphVT3Y 00:39:25.740 19:35:38 keyring_file -- keyring/file.sh@109 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.tw5Rv6bC7R 00:39:25.740 19:35:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.tw5Rv6bC7R 00:39:25.740 19:35:38 keyring_file -- keyring/file.sh@110 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:25.740 19:35:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:39:25.999 nvme0n1 00:39:25.999 19:35:38 keyring_file -- keyring/file.sh@113 -- # bperf_cmd save_config 00:39:25.999 19:35:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock save_config 00:39:26.259 19:35:38 keyring_file -- keyring/file.sh@113 -- # config='{ 00:39:26.259 "subsystems": [ 00:39:26.259 { 00:39:26.259 "subsystem": "keyring", 00:39:26.259 "config": [ 00:39:26.259 { 00:39:26.259 "method": "keyring_file_add_key", 00:39:26.259 "params": { 00:39:26.259 "name": "key0", 00:39:26.259 "path": "/tmp/tmp.BkUmphVT3Y" 00:39:26.259 } 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "method": "keyring_file_add_key", 00:39:26.259 "params": { 00:39:26.259 "name": "key1", 00:39:26.259 "path": "/tmp/tmp.tw5Rv6bC7R" 00:39:26.259 } 00:39:26.259 } 00:39:26.259 ] 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "subsystem": "iobuf", 00:39:26.259 "config": [ 00:39:26.259 { 00:39:26.259 "method": "iobuf_set_options", 00:39:26.259 "params": { 00:39:26.259 "small_pool_count": 8192, 00:39:26.259 "large_pool_count": 1024, 00:39:26.259 "small_bufsize": 8192, 00:39:26.259 "large_bufsize": 135168, 00:39:26.259 "enable_numa": false 00:39:26.259 } 00:39:26.259 } 00:39:26.259 ] 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "subsystem": "sock", 00:39:26.259 "config": [ 00:39:26.259 { 00:39:26.259 "method": "sock_set_default_impl", 00:39:26.259 "params": { 00:39:26.259 "impl_name": "posix" 00:39:26.259 } 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "method": "sock_impl_set_options", 00:39:26.259 "params": { 00:39:26.259 "impl_name": "ssl", 00:39:26.259 "recv_buf_size": 4096, 00:39:26.259 "send_buf_size": 4096, 00:39:26.259 "enable_recv_pipe": true, 00:39:26.259 "enable_quickack": false, 00:39:26.259 "enable_placement_id": 0, 00:39:26.259 "enable_zerocopy_send_server": true, 00:39:26.259 "enable_zerocopy_send_client": false, 00:39:26.259 "zerocopy_threshold": 0, 00:39:26.259 "tls_version": 0, 00:39:26.259 "enable_ktls": false 00:39:26.259 } 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "method": "sock_impl_set_options", 00:39:26.259 "params": { 00:39:26.259 "impl_name": "posix", 00:39:26.259 "recv_buf_size": 2097152, 00:39:26.259 "send_buf_size": 2097152, 00:39:26.259 "enable_recv_pipe": true, 00:39:26.259 "enable_quickack": false, 00:39:26.259 "enable_placement_id": 0, 00:39:26.259 "enable_zerocopy_send_server": true, 00:39:26.259 "enable_zerocopy_send_client": false, 00:39:26.259 "zerocopy_threshold": 0, 00:39:26.259 "tls_version": 0, 00:39:26.259 "enable_ktls": false 00:39:26.259 } 00:39:26.259 } 00:39:26.259 ] 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "subsystem": "vmd", 00:39:26.259 "config": [] 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "subsystem": "accel", 00:39:26.259 "config": [ 00:39:26.259 { 00:39:26.259 "method": "accel_set_options", 00:39:26.259 "params": { 00:39:26.259 "small_cache_size": 128, 00:39:26.259 "large_cache_size": 16, 00:39:26.259 "task_count": 2048, 00:39:26.259 "sequence_count": 2048, 00:39:26.259 "buf_count": 2048 00:39:26.259 } 00:39:26.259 } 00:39:26.259 ] 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "subsystem": "bdev", 00:39:26.259 "config": [ 00:39:26.259 { 00:39:26.259 "method": "bdev_set_options", 00:39:26.259 "params": { 00:39:26.259 "bdev_io_pool_size": 65535, 00:39:26.259 "bdev_io_cache_size": 256, 00:39:26.259 "bdev_auto_examine": true, 00:39:26.259 "iobuf_small_cache_size": 128, 00:39:26.259 "iobuf_large_cache_size": 16 00:39:26.259 } 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "method": "bdev_raid_set_options", 00:39:26.259 "params": { 00:39:26.259 "process_window_size_kb": 1024, 00:39:26.259 "process_max_bandwidth_mb_sec": 0 00:39:26.259 } 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "method": "bdev_iscsi_set_options", 00:39:26.259 "params": { 00:39:26.259 "timeout_sec": 30 00:39:26.259 } 00:39:26.259 }, 00:39:26.259 { 00:39:26.259 "method": "bdev_nvme_set_options", 00:39:26.259 "params": { 00:39:26.259 "action_on_timeout": "none", 00:39:26.259 "timeout_us": 0, 00:39:26.259 "timeout_admin_us": 0, 00:39:26.259 "keep_alive_timeout_ms": 10000, 00:39:26.259 "arbitration_burst": 0, 00:39:26.259 "low_priority_weight": 0, 00:39:26.259 "medium_priority_weight": 0, 00:39:26.259 "high_priority_weight": 0, 00:39:26.259 "nvme_adminq_poll_period_us": 10000, 00:39:26.259 "nvme_ioq_poll_period_us": 0, 00:39:26.259 "io_queue_requests": 512, 00:39:26.259 "delay_cmd_submit": true, 00:39:26.259 "transport_retry_count": 4, 00:39:26.259 "bdev_retry_count": 3, 00:39:26.259 "transport_ack_timeout": 0, 00:39:26.259 "ctrlr_loss_timeout_sec": 0, 00:39:26.259 "reconnect_delay_sec": 0, 00:39:26.259 "fast_io_fail_timeout_sec": 0, 00:39:26.259 "disable_auto_failback": false, 00:39:26.259 "generate_uuids": false, 00:39:26.259 "transport_tos": 0, 00:39:26.259 "nvme_error_stat": false, 00:39:26.259 "rdma_srq_size": 0, 00:39:26.259 "io_path_stat": false, 00:39:26.259 "allow_accel_sequence": false, 00:39:26.259 "rdma_max_cq_size": 0, 00:39:26.259 "rdma_cm_event_timeout_ms": 0, 00:39:26.259 "dhchap_digests": [ 00:39:26.259 "sha256", 00:39:26.260 "sha384", 00:39:26.260 "sha512" 00:39:26.260 ], 00:39:26.260 "dhchap_dhgroups": [ 00:39:26.260 "null", 00:39:26.260 "ffdhe2048", 00:39:26.260 "ffdhe3072", 00:39:26.260 "ffdhe4096", 00:39:26.260 "ffdhe6144", 00:39:26.260 "ffdhe8192" 00:39:26.260 ] 00:39:26.260 } 00:39:26.260 }, 00:39:26.260 { 00:39:26.260 "method": "bdev_nvme_attach_controller", 00:39:26.260 "params": { 00:39:26.260 "name": "nvme0", 00:39:26.260 "trtype": "TCP", 00:39:26.260 "adrfam": "IPv4", 00:39:26.260 "traddr": "127.0.0.1", 00:39:26.260 "trsvcid": "4420", 00:39:26.260 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:26.260 "prchk_reftag": false, 00:39:26.260 "prchk_guard": false, 00:39:26.260 "ctrlr_loss_timeout_sec": 0, 00:39:26.260 "reconnect_delay_sec": 0, 00:39:26.260 "fast_io_fail_timeout_sec": 0, 00:39:26.260 "psk": "key0", 00:39:26.260 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:26.260 "hdgst": false, 00:39:26.260 "ddgst": false, 00:39:26.260 "multipath": "multipath" 00:39:26.260 } 00:39:26.260 }, 00:39:26.260 { 00:39:26.260 "method": "bdev_nvme_set_hotplug", 00:39:26.260 "params": { 00:39:26.260 "period_us": 100000, 00:39:26.260 "enable": false 00:39:26.260 } 00:39:26.260 }, 00:39:26.260 { 00:39:26.260 "method": "bdev_wait_for_examine" 00:39:26.260 } 00:39:26.260 ] 00:39:26.260 }, 00:39:26.260 { 00:39:26.260 "subsystem": "nbd", 00:39:26.260 "config": [] 00:39:26.260 } 00:39:26.260 ] 00:39:26.260 }' 00:39:26.260 19:35:38 keyring_file -- keyring/file.sh@115 -- # killprocess 24042 00:39:26.260 19:35:38 keyring_file -- common/autotest_common.sh@954 -- # '[' -z 24042 ']' 00:39:26.260 19:35:38 keyring_file -- common/autotest_common.sh@958 -- # kill -0 24042 00:39:26.260 19:35:38 keyring_file -- common/autotest_common.sh@959 -- # uname 00:39:26.260 19:35:38 keyring_file -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:39:26.260 19:35:38 keyring_file -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 24042 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@972 -- # echo 'killing process with pid 24042' 00:39:26.522 killing process with pid 24042 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@973 -- # kill 24042 00:39:26.522 Received shutdown signal, test time was about 1.000000 seconds 00:39:26.522 00:39:26.522 Latency(us) 00:39:26.522 [2024-11-26T18:35:39.147Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:39:26.522 [2024-11-26T18:35:39.147Z] =================================================================================================================== 00:39:26.522 [2024-11-26T18:35:39.147Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@978 -- # wait 24042 00:39:26.522 19:35:38 keyring_file -- keyring/file.sh@118 -- # bperfpid=25842 00:39:26.522 19:35:38 keyring_file -- keyring/file.sh@120 -- # waitforlisten 25842 /var/tmp/bperf.sock 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@835 -- # '[' -z 25842 ']' 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:39:26.522 19:35:38 keyring_file -- keyring/file.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z -c /dev/fd/63 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@840 -- # local max_retries=100 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:39:26.522 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@844 -- # xtrace_disable 00:39:26.522 19:35:38 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:39:26.522 19:35:38 keyring_file -- keyring/file.sh@116 -- # echo '{ 00:39:26.522 "subsystems": [ 00:39:26.522 { 00:39:26.522 "subsystem": "keyring", 00:39:26.522 "config": [ 00:39:26.522 { 00:39:26.522 "method": "keyring_file_add_key", 00:39:26.522 "params": { 00:39:26.522 "name": "key0", 00:39:26.522 "path": "/tmp/tmp.BkUmphVT3Y" 00:39:26.522 } 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "method": "keyring_file_add_key", 00:39:26.522 "params": { 00:39:26.522 "name": "key1", 00:39:26.522 "path": "/tmp/tmp.tw5Rv6bC7R" 00:39:26.522 } 00:39:26.522 } 00:39:26.522 ] 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "subsystem": "iobuf", 00:39:26.522 "config": [ 00:39:26.522 { 00:39:26.522 "method": "iobuf_set_options", 00:39:26.522 "params": { 00:39:26.522 "small_pool_count": 8192, 00:39:26.522 "large_pool_count": 1024, 00:39:26.522 "small_bufsize": 8192, 00:39:26.522 "large_bufsize": 135168, 00:39:26.522 "enable_numa": false 00:39:26.522 } 00:39:26.522 } 00:39:26.522 ] 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "subsystem": "sock", 00:39:26.522 "config": [ 00:39:26.522 { 00:39:26.522 "method": "sock_set_default_impl", 00:39:26.522 "params": { 00:39:26.522 "impl_name": "posix" 00:39:26.522 } 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "method": "sock_impl_set_options", 00:39:26.522 "params": { 00:39:26.522 "impl_name": "ssl", 00:39:26.522 "recv_buf_size": 4096, 00:39:26.522 "send_buf_size": 4096, 00:39:26.522 "enable_recv_pipe": true, 00:39:26.522 "enable_quickack": false, 00:39:26.522 "enable_placement_id": 0, 00:39:26.522 "enable_zerocopy_send_server": true, 00:39:26.522 "enable_zerocopy_send_client": false, 00:39:26.522 "zerocopy_threshold": 0, 00:39:26.522 "tls_version": 0, 00:39:26.522 "enable_ktls": false 00:39:26.522 } 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "method": "sock_impl_set_options", 00:39:26.522 "params": { 00:39:26.522 "impl_name": "posix", 00:39:26.522 "recv_buf_size": 2097152, 00:39:26.522 "send_buf_size": 2097152, 00:39:26.522 "enable_recv_pipe": true, 00:39:26.522 "enable_quickack": false, 00:39:26.522 "enable_placement_id": 0, 00:39:26.522 "enable_zerocopy_send_server": true, 00:39:26.522 "enable_zerocopy_send_client": false, 00:39:26.522 "zerocopy_threshold": 0, 00:39:26.522 "tls_version": 0, 00:39:26.522 "enable_ktls": false 00:39:26.522 } 00:39:26.522 } 00:39:26.522 ] 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "subsystem": "vmd", 00:39:26.522 "config": [] 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "subsystem": "accel", 00:39:26.522 "config": [ 00:39:26.522 { 00:39:26.522 "method": "accel_set_options", 00:39:26.522 "params": { 00:39:26.522 "small_cache_size": 128, 00:39:26.522 "large_cache_size": 16, 00:39:26.522 "task_count": 2048, 00:39:26.522 "sequence_count": 2048, 00:39:26.522 "buf_count": 2048 00:39:26.522 } 00:39:26.522 } 00:39:26.522 ] 00:39:26.522 }, 00:39:26.522 { 00:39:26.522 "subsystem": "bdev", 00:39:26.522 "config": [ 00:39:26.522 { 00:39:26.522 "method": "bdev_set_options", 00:39:26.522 "params": { 00:39:26.522 "bdev_io_pool_size": 65535, 00:39:26.522 "bdev_io_cache_size": 256, 00:39:26.522 "bdev_auto_examine": true, 00:39:26.522 "iobuf_small_cache_size": 128, 00:39:26.522 "iobuf_large_cache_size": 16 00:39:26.522 } 00:39:26.523 }, 00:39:26.523 { 00:39:26.523 "method": "bdev_raid_set_options", 00:39:26.523 "params": { 00:39:26.523 "process_window_size_kb": 1024, 00:39:26.523 "process_max_bandwidth_mb_sec": 0 00:39:26.523 } 00:39:26.523 }, 00:39:26.523 { 00:39:26.523 "method": "bdev_iscsi_set_options", 00:39:26.523 "params": { 00:39:26.523 "timeout_sec": 30 00:39:26.523 } 00:39:26.523 }, 00:39:26.523 { 00:39:26.523 "method": "bdev_nvme_set_options", 00:39:26.523 "params": { 00:39:26.523 "action_on_timeout": "none", 00:39:26.523 "timeout_us": 0, 00:39:26.523 "timeout_admin_us": 0, 00:39:26.523 "keep_alive_timeout_ms": 10000, 00:39:26.523 "arbitration_burst": 0, 00:39:26.523 "low_priority_weight": 0, 00:39:26.523 "medium_priority_weight": 0, 00:39:26.523 "high_priority_weight": 0, 00:39:26.523 "nvme_adminq_poll_period_us": 10000, 00:39:26.523 "nvme_ioq_poll_period_us": 0, 00:39:26.523 "io_queue_requests": 512, 00:39:26.523 "delay_cmd_submit": true, 00:39:26.523 "transport_retry_count": 4, 00:39:26.523 "bdev_retry_count": 3, 00:39:26.523 "transport_ack_timeout": 0, 00:39:26.523 "ctrlr_loss_timeout_sec": 0, 00:39:26.523 "reconnect_delay_sec": 0, 00:39:26.523 "fast_io_fail_timeout_sec": 0, 00:39:26.523 "disable_auto_failback": false, 00:39:26.523 "generate_uuids": false, 00:39:26.523 "transport_tos": 0, 00:39:26.523 "nvme_error_stat": false, 00:39:26.523 "rdma_srq_size": 0, 00:39:26.523 "io_path_stat": false, 00:39:26.523 "allow_accel_sequence": false, 00:39:26.523 "rdma_max_cq_size": 0, 00:39:26.523 "rdma_cm_event_timeout_ms": 0, 00:39:26.523 "dhchap_digests": [ 00:39:26.523 "sha256", 00:39:26.523 "sha384", 00:39:26.523 "sha512" 00:39:26.523 ], 00:39:26.523 "dhchap_dhgroups": [ 00:39:26.523 "null", 00:39:26.523 "ffdhe2048", 00:39:26.523 "ffdhe3072", 00:39:26.523 "ffdhe4096", 00:39:26.523 "ffdhe6144", 00:39:26.523 "ffdhe8192" 00:39:26.523 ] 00:39:26.523 } 00:39:26.523 }, 00:39:26.523 { 00:39:26.523 "method": "bdev_nvme_attach_controller", 00:39:26.523 "params": { 00:39:26.523 "name": "nvme0", 00:39:26.523 "trtype": "TCP", 00:39:26.523 "adrfam": "IPv4", 00:39:26.523 "traddr": "127.0.0.1", 00:39:26.523 "trsvcid": "4420", 00:39:26.523 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:26.523 "prchk_reftag": false, 00:39:26.523 "prchk_guard": false, 00:39:26.523 "ctrlr_loss_timeout_sec": 0, 00:39:26.523 "reconnect_delay_sec": 0, 00:39:26.523 "fast_io_fail_timeout_sec": 0, 00:39:26.523 "psk": "key0", 00:39:26.523 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:26.523 "hdgst": false, 00:39:26.523 "ddgst": false, 00:39:26.523 "multipath": "multipath" 00:39:26.523 } 00:39:26.523 }, 00:39:26.523 { 00:39:26.523 "method": "bdev_nvme_set_hotplug", 00:39:26.523 "params": { 00:39:26.523 "period_us": 100000, 00:39:26.523 "enable": false 00:39:26.523 } 00:39:26.523 }, 00:39:26.523 { 00:39:26.523 "method": "bdev_wait_for_examine" 00:39:26.523 } 00:39:26.523 ] 00:39:26.523 }, 00:39:26.523 { 00:39:26.523 "subsystem": "nbd", 00:39:26.523 "config": [] 00:39:26.523 } 00:39:26.523 ] 00:39:26.523 }' 00:39:26.523 [2024-11-26 19:35:39.044303] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:39:26.523 [2024-11-26 19:35:39.044356] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid25842 ] 00:39:26.523 [2024-11-26 19:35:39.134601] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:39:26.783 [2024-11-26 19:35:39.163264] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:39:26.783 [2024-11-26 19:35:39.307906] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:39:27.399 19:35:39 keyring_file -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:39:27.399 19:35:39 keyring_file -- common/autotest_common.sh@868 -- # return 0 00:39:27.399 19:35:39 keyring_file -- keyring/file.sh@121 -- # bperf_cmd keyring_get_keys 00:39:27.399 19:35:39 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:27.399 19:35:39 keyring_file -- keyring/file.sh@121 -- # jq length 00:39:27.692 19:35:40 keyring_file -- keyring/file.sh@121 -- # (( 2 == 2 )) 00:39:27.692 19:35:40 keyring_file -- keyring/file.sh@122 -- # get_refcnt key0 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:27.692 19:35:40 keyring_file -- keyring/file.sh@122 -- # (( 2 == 2 )) 00:39:27.692 19:35:40 keyring_file -- keyring/file.sh@123 -- # get_refcnt key1 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:39:27.692 19:35:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:27.978 19:35:40 keyring_file -- keyring/file.sh@123 -- # (( 1 == 1 )) 00:39:27.978 19:35:40 keyring_file -- keyring/file.sh@124 -- # bperf_cmd bdev_nvme_get_controllers 00:39:27.978 19:35:40 keyring_file -- keyring/file.sh@124 -- # jq -r '.[].name' 00:39:27.978 19:35:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_get_controllers 00:39:27.978 19:35:40 keyring_file -- keyring/file.sh@124 -- # [[ nvme0 == nvme0 ]] 00:39:27.978 19:35:40 keyring_file -- keyring/file.sh@1 -- # cleanup 00:39:27.978 19:35:40 keyring_file -- keyring/file.sh@19 -- # rm -f /tmp/tmp.BkUmphVT3Y /tmp/tmp.tw5Rv6bC7R 00:39:27.978 19:35:40 keyring_file -- keyring/file.sh@20 -- # killprocess 25842 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@954 -- # '[' -z 25842 ']' 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@958 -- # kill -0 25842 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@959 -- # uname 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 25842 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:39:27.978 19:35:40 keyring_file -- common/autotest_common.sh@972 -- # echo 'killing process with pid 25842' 00:39:27.978 killing process with pid 25842 00:39:27.979 19:35:40 keyring_file -- common/autotest_common.sh@973 -- # kill 25842 00:39:27.979 Received shutdown signal, test time was about 1.000000 seconds 00:39:27.979 00:39:27.979 Latency(us) 00:39:27.979 [2024-11-26T18:35:40.604Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:39:27.979 [2024-11-26T18:35:40.604Z] =================================================================================================================== 00:39:27.979 [2024-11-26T18:35:40.604Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@978 -- # wait 25842 00:39:28.240 19:35:40 keyring_file -- keyring/file.sh@21 -- # killprocess 24021 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@954 -- # '[' -z 24021 ']' 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@958 -- # kill -0 24021 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@959 -- # uname 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 24021 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@972 -- # echo 'killing process with pid 24021' 00:39:28.240 killing process with pid 24021 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@973 -- # kill 24021 00:39:28.240 19:35:40 keyring_file -- common/autotest_common.sh@978 -- # wait 24021 00:39:28.501 00:39:28.501 real 0m11.673s 00:39:28.501 user 0m28.007s 00:39:28.501 sys 0m2.637s 00:39:28.501 19:35:40 keyring_file -- common/autotest_common.sh@1130 -- # xtrace_disable 00:39:28.501 19:35:40 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:39:28.501 ************************************ 00:39:28.501 END TEST keyring_file 00:39:28.501 ************************************ 00:39:28.501 19:35:41 -- spdk/autotest.sh@293 -- # [[ y == y ]] 00:39:28.501 19:35:41 -- spdk/autotest.sh@294 -- # run_test keyring_linux /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:39:28.501 19:35:41 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:39:28.501 19:35:41 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:39:28.501 19:35:41 -- common/autotest_common.sh@10 -- # set +x 00:39:28.502 ************************************ 00:39:28.502 START TEST keyring_linux 00:39:28.502 ************************************ 00:39:28.502 19:35:41 keyring_linux -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:39:28.502 Joined session keyring: 360534669 00:39:28.763 * Looking for test storage... 00:39:28.763 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:39:28.763 19:35:41 keyring_linux -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:39:28.763 19:35:41 keyring_linux -- common/autotest_common.sh@1693 -- # lcov --version 00:39:28.763 19:35:41 keyring_linux -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:39:28.763 19:35:41 keyring_linux -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@333 -- # local ver1 ver1_l 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@334 -- # local ver2 ver2_l 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@336 -- # IFS=.-: 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@336 -- # read -ra ver1 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@337 -- # IFS=.-: 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@337 -- # read -ra ver2 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@338 -- # local 'op=<' 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@340 -- # ver1_l=2 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@341 -- # ver2_l=1 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@344 -- # case "$op" in 00:39:28.763 19:35:41 keyring_linux -- scripts/common.sh@345 -- # : 1 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@364 -- # (( v = 0 )) 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@365 -- # decimal 1 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@353 -- # local d=1 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@355 -- # echo 1 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@365 -- # ver1[v]=1 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@366 -- # decimal 2 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@353 -- # local d=2 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@355 -- # echo 2 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@366 -- # ver2[v]=2 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@368 -- # return 0 00:39:28.764 19:35:41 keyring_linux -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:39:28.764 19:35:41 keyring_linux -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:39:28.764 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:28.764 --rc genhtml_branch_coverage=1 00:39:28.764 --rc genhtml_function_coverage=1 00:39:28.764 --rc genhtml_legend=1 00:39:28.764 --rc geninfo_all_blocks=1 00:39:28.764 --rc geninfo_unexecuted_blocks=1 00:39:28.764 00:39:28.764 ' 00:39:28.764 19:35:41 keyring_linux -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:39:28.764 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:28.764 --rc genhtml_branch_coverage=1 00:39:28.764 --rc genhtml_function_coverage=1 00:39:28.764 --rc genhtml_legend=1 00:39:28.764 --rc geninfo_all_blocks=1 00:39:28.764 --rc geninfo_unexecuted_blocks=1 00:39:28.764 00:39:28.764 ' 00:39:28.764 19:35:41 keyring_linux -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:39:28.764 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:28.764 --rc genhtml_branch_coverage=1 00:39:28.764 --rc genhtml_function_coverage=1 00:39:28.764 --rc genhtml_legend=1 00:39:28.764 --rc geninfo_all_blocks=1 00:39:28.764 --rc geninfo_unexecuted_blocks=1 00:39:28.764 00:39:28.764 ' 00:39:28.764 19:35:41 keyring_linux -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:39:28.764 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:28.764 --rc genhtml_branch_coverage=1 00:39:28.764 --rc genhtml_function_coverage=1 00:39:28.764 --rc genhtml_legend=1 00:39:28.764 --rc geninfo_all_blocks=1 00:39:28.764 --rc geninfo_unexecuted_blocks=1 00:39:28.764 00:39:28.764 ' 00:39:28.764 19:35:41 keyring_linux -- keyring/linux.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@7 -- # uname -s 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00539ede-7deb-ec11-9bc7-a4bf01928396 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@18 -- # NVME_HOSTID=00539ede-7deb-ec11-9bc7-a4bf01928396 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@15 -- # shopt -s extglob 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:39:28.764 19:35:41 keyring_linux -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:39:28.764 19:35:41 keyring_linux -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:28.764 19:35:41 keyring_linux -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:28.764 19:35:41 keyring_linux -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:28.764 19:35:41 keyring_linux -- paths/export.sh@5 -- # export PATH 00:39:28.764 19:35:41 keyring_linux -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@51 -- # : 0 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:39:28.764 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@55 -- # have_pci_nics=0 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:39:28.764 19:35:41 keyring_linux -- keyring/linux.sh@11 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:39:28.764 19:35:41 keyring_linux -- keyring/linux.sh@12 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:39:28.764 19:35:41 keyring_linux -- keyring/linux.sh@13 -- # key0=00112233445566778899aabbccddeeff 00:39:28.764 19:35:41 keyring_linux -- keyring/linux.sh@14 -- # key1=112233445566778899aabbccddeeff00 00:39:28.764 19:35:41 keyring_linux -- keyring/linux.sh@45 -- # trap cleanup EXIT 00:39:28.764 19:35:41 keyring_linux -- keyring/linux.sh@47 -- # prep_key key0 00112233445566778899aabbccddeeff 0 /tmp/:spdk-test:key0 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@17 -- # name=key0 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key0 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@730 -- # local prefix key digest 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@732 -- # digest=0 00:39:28.764 19:35:41 keyring_linux -- nvmf/common.sh@733 -- # python - 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key0 00:39:28.764 19:35:41 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key0 00:39:28.764 /tmp/:spdk-test:key0 00:39:28.765 19:35:41 keyring_linux -- keyring/linux.sh@48 -- # prep_key key1 112233445566778899aabbccddeeff00 0 /tmp/:spdk-test:key1 00:39:28.765 19:35:41 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:39:28.765 19:35:41 keyring_linux -- keyring/common.sh@17 -- # name=key1 00:39:28.765 19:35:41 keyring_linux -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:39:28.765 19:35:41 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:39:28.765 19:35:41 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key1 00:39:28.765 19:35:41 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:39:28.765 19:35:41 keyring_linux -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:39:28.765 19:35:41 keyring_linux -- nvmf/common.sh@730 -- # local prefix key digest 00:39:28.765 19:35:41 keyring_linux -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:39:28.765 19:35:41 keyring_linux -- nvmf/common.sh@732 -- # key=112233445566778899aabbccddeeff00 00:39:28.765 19:35:41 keyring_linux -- nvmf/common.sh@732 -- # digest=0 00:39:28.765 19:35:41 keyring_linux -- nvmf/common.sh@733 -- # python - 00:39:29.027 19:35:41 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key1 00:39:29.027 19:35:41 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key1 00:39:29.027 /tmp/:spdk-test:key1 00:39:29.027 19:35:41 keyring_linux -- keyring/linux.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:39:29.027 19:35:41 keyring_linux -- keyring/linux.sh@51 -- # tgtpid=26285 00:39:29.027 19:35:41 keyring_linux -- keyring/linux.sh@53 -- # waitforlisten 26285 00:39:29.027 19:35:41 keyring_linux -- common/autotest_common.sh@835 -- # '[' -z 26285 ']' 00:39:29.027 19:35:41 keyring_linux -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:39:29.027 19:35:41 keyring_linux -- common/autotest_common.sh@840 -- # local max_retries=100 00:39:29.027 19:35:41 keyring_linux -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:39:29.027 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:39:29.027 19:35:41 keyring_linux -- common/autotest_common.sh@844 -- # xtrace_disable 00:39:29.027 19:35:41 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:39:29.027 [2024-11-26 19:35:41.439613] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:39:29.027 [2024-11-26 19:35:41.439669] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid26285 ] 00:39:29.027 [2024-11-26 19:35:41.520577] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:39:29.027 [2024-11-26 19:35:41.558824] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@868 -- # return 0 00:39:29.971 19:35:42 keyring_linux -- keyring/linux.sh@54 -- # rpc_cmd 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:39:29.971 [2024-11-26 19:35:42.254662] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:39:29.971 null0 00:39:29.971 [2024-11-26 19:35:42.286701] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:39:29.971 [2024-11-26 19:35:42.287117] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:29.971 19:35:42 keyring_linux -- keyring/linux.sh@66 -- # keyctl add user :spdk-test:key0 NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: @s 00:39:29.971 590974616 00:39:29.971 19:35:42 keyring_linux -- keyring/linux.sh@67 -- # keyctl add user :spdk-test:key1 NVMeTLSkey-1:00:MTEyMjMzNDQ1NTY2Nzc4ODk5YWFiYmNjZGRlZWZmMDA6CPcs: @s 00:39:29.971 482080255 00:39:29.971 19:35:42 keyring_linux -- keyring/linux.sh@70 -- # bperfpid=26621 00:39:29.971 19:35:42 keyring_linux -- keyring/linux.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randread -t 1 -m 2 -r /var/tmp/bperf.sock -z --wait-for-rpc 00:39:29.971 19:35:42 keyring_linux -- keyring/linux.sh@72 -- # waitforlisten 26621 /var/tmp/bperf.sock 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@835 -- # '[' -z 26621 ']' 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@840 -- # local max_retries=100 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:39:29.971 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@844 -- # xtrace_disable 00:39:29.971 19:35:42 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:39:29.971 [2024-11-26 19:35:42.364201] Starting SPDK v25.01-pre git sha1 bb877d8c1 / DPDK 24.03.0 initialization... 00:39:29.971 [2024-11-26 19:35:42.364249] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid26621 ] 00:39:29.971 [2024-11-26 19:35:42.453723] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:39:29.971 [2024-11-26 19:35:42.483640] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:39:30.542 19:35:43 keyring_linux -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:39:30.542 19:35:43 keyring_linux -- common/autotest_common.sh@868 -- # return 0 00:39:30.542 19:35:43 keyring_linux -- keyring/linux.sh@73 -- # bperf_cmd keyring_linux_set_options --enable 00:39:30.542 19:35:43 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_linux_set_options --enable 00:39:30.803 19:35:43 keyring_linux -- keyring/linux.sh@74 -- # bperf_cmd framework_start_init 00:39:30.803 19:35:43 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:39:31.064 19:35:43 keyring_linux -- keyring/linux.sh@75 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:39:31.064 19:35:43 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:39:31.064 [2024-11-26 19:35:43.685188] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:39:31.325 nvme0n1 00:39:31.325 19:35:43 keyring_linux -- keyring/linux.sh@77 -- # check_keys 1 :spdk-test:key0 00:39:31.325 19:35:43 keyring_linux -- keyring/linux.sh@19 -- # local count=1 name=:spdk-test:key0 00:39:31.325 19:35:43 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:39:31.325 19:35:43 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:39:31.325 19:35:43 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:39:31.325 19:35:43 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:31.586 19:35:43 keyring_linux -- keyring/linux.sh@22 -- # (( 1 == count )) 00:39:31.586 19:35:43 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:39:31.586 19:35:43 keyring_linux -- keyring/linux.sh@25 -- # get_key :spdk-test:key0 00:39:31.586 19:35:43 keyring_linux -- keyring/linux.sh@25 -- # jq -r .sn 00:39:31.586 19:35:43 keyring_linux -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:39:31.586 19:35:43 keyring_linux -- keyring/common.sh@10 -- # jq '.[] | select(.name == ":spdk-test:key0")' 00:39:31.586 19:35:43 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:31.586 19:35:44 keyring_linux -- keyring/linux.sh@25 -- # sn=590974616 00:39:31.586 19:35:44 keyring_linux -- keyring/linux.sh@26 -- # get_keysn :spdk-test:key0 00:39:31.586 19:35:44 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:39:31.586 19:35:44 keyring_linux -- keyring/linux.sh@26 -- # [[ 590974616 == \5\9\0\9\7\4\6\1\6 ]] 00:39:31.586 19:35:44 keyring_linux -- keyring/linux.sh@27 -- # keyctl print 590974616 00:39:31.586 19:35:44 keyring_linux -- keyring/linux.sh@27 -- # [[ NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: == \N\V\M\e\T\L\S\k\e\y\-\1\:\0\0\:\M\D\A\x\M\T\I\y\M\z\M\0\N\D\U\1\N\j\Y\3\N\z\g\4\O\T\l\h\Y\W\J\i\Y\2\N\k\Z\G\V\l\Z\m\Z\w\J\E\i\Q\: ]] 00:39:31.586 19:35:44 keyring_linux -- keyring/linux.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:39:31.848 Running I/O for 1 seconds... 00:39:32.791 16000.00 IOPS, 62.50 MiB/s 00:39:32.791 Latency(us) 00:39:32.791 [2024-11-26T18:35:45.416Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:39:32.791 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:39:32.791 nvme0n1 : 1.01 16002.37 62.51 0.00 0.00 7964.03 6690.13 14636.37 00:39:32.791 [2024-11-26T18:35:45.416Z] =================================================================================================================== 00:39:32.791 [2024-11-26T18:35:45.416Z] Total : 16002.37 62.51 0.00 0.00 7964.03 6690.13 14636.37 00:39:32.791 { 00:39:32.791 "results": [ 00:39:32.791 { 00:39:32.791 "job": "nvme0n1", 00:39:32.791 "core_mask": "0x2", 00:39:32.791 "workload": "randread", 00:39:32.791 "status": "finished", 00:39:32.791 "queue_depth": 128, 00:39:32.791 "io_size": 4096, 00:39:32.791 "runtime": 1.007913, 00:39:32.791 "iops": 16002.373220704565, 00:39:32.791 "mibps": 62.50927039337721, 00:39:32.791 "io_failed": 0, 00:39:32.791 "io_timeout": 0, 00:39:32.791 "avg_latency_us": 7964.029373178745, 00:39:32.791 "min_latency_us": 6690.133333333333, 00:39:32.791 "max_latency_us": 14636.373333333333 00:39:32.791 } 00:39:32.791 ], 00:39:32.791 "core_count": 1 00:39:32.791 } 00:39:32.791 19:35:45 keyring_linux -- keyring/linux.sh@80 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:39:32.791 19:35:45 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@81 -- # check_keys 0 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@19 -- # local count=0 name= 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:39:33.052 19:35:45 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@22 -- # (( 0 == count )) 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@23 -- # return 00:39:33.052 19:35:45 keyring_linux -- keyring/linux.sh@84 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:39:33.052 19:35:45 keyring_linux -- common/autotest_common.sh@652 -- # local es=0 00:39:33.052 19:35:45 keyring_linux -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:39:33.052 19:35:45 keyring_linux -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:39:33.052 19:35:45 keyring_linux -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:33.052 19:35:45 keyring_linux -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:39:33.052 19:35:45 keyring_linux -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:33.052 19:35:45 keyring_linux -- common/autotest_common.sh@655 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:39:33.052 19:35:45 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:39:33.313 [2024-11-26 19:35:45.758305] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:39:33.313 [2024-11-26 19:35:45.759050] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd8b7a0 (107): Transport endpoint is not connected 00:39:33.313 [2024-11-26 19:35:45.760046] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd8b7a0 (9): Bad file descriptor 00:39:33.313 [2024-11-26 19:35:45.761049] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] Ctrlr is in error state 00:39:33.313 [2024-11-26 19:35:45.761062] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:39:33.313 [2024-11-26 19:35:45.761068] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:39:33.313 [2024-11-26 19:35:45.761075] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] in failed state. 00:39:33.313 request: 00:39:33.313 { 00:39:33.313 "name": "nvme0", 00:39:33.313 "trtype": "tcp", 00:39:33.313 "traddr": "127.0.0.1", 00:39:33.313 "adrfam": "ipv4", 00:39:33.313 "trsvcid": "4420", 00:39:33.313 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:33.313 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:33.313 "prchk_reftag": false, 00:39:33.313 "prchk_guard": false, 00:39:33.313 "hdgst": false, 00:39:33.313 "ddgst": false, 00:39:33.313 "psk": ":spdk-test:key1", 00:39:33.313 "allow_unrecognized_csi": false, 00:39:33.313 "method": "bdev_nvme_attach_controller", 00:39:33.313 "req_id": 1 00:39:33.313 } 00:39:33.313 Got JSON-RPC error response 00:39:33.313 response: 00:39:33.313 { 00:39:33.313 "code": -5, 00:39:33.313 "message": "Input/output error" 00:39:33.313 } 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@655 -- # es=1 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@1 -- # cleanup 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key0 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@31 -- # local name=key0 sn 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key0 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@33 -- # sn=590974616 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 590974616 00:39:33.313 1 links removed 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key1 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@31 -- # local name=key1 sn 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key1 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key1 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@33 -- # sn=482080255 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 482080255 00:39:33.313 1 links removed 00:39:33.313 19:35:45 keyring_linux -- keyring/linux.sh@41 -- # killprocess 26621 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@954 -- # '[' -z 26621 ']' 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@958 -- # kill -0 26621 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@959 -- # uname 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 26621 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@972 -- # echo 'killing process with pid 26621' 00:39:33.313 killing process with pid 26621 00:39:33.313 19:35:45 keyring_linux -- common/autotest_common.sh@973 -- # kill 26621 00:39:33.313 Received shutdown signal, test time was about 1.000000 seconds 00:39:33.313 00:39:33.313 Latency(us) 00:39:33.314 [2024-11-26T18:35:45.939Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:39:33.314 [2024-11-26T18:35:45.939Z] =================================================================================================================== 00:39:33.314 [2024-11-26T18:35:45.939Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:39:33.314 19:35:45 keyring_linux -- common/autotest_common.sh@978 -- # wait 26621 00:39:33.575 19:35:45 keyring_linux -- keyring/linux.sh@42 -- # killprocess 26285 00:39:33.575 19:35:45 keyring_linux -- common/autotest_common.sh@954 -- # '[' -z 26285 ']' 00:39:33.575 19:35:45 keyring_linux -- common/autotest_common.sh@958 -- # kill -0 26285 00:39:33.575 19:35:45 keyring_linux -- common/autotest_common.sh@959 -- # uname 00:39:33.575 19:35:45 keyring_linux -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:39:33.575 19:35:45 keyring_linux -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 26285 00:39:33.575 19:35:46 keyring_linux -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:39:33.575 19:35:46 keyring_linux -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:39:33.575 19:35:46 keyring_linux -- common/autotest_common.sh@972 -- # echo 'killing process with pid 26285' 00:39:33.575 killing process with pid 26285 00:39:33.575 19:35:46 keyring_linux -- common/autotest_common.sh@973 -- # kill 26285 00:39:33.575 19:35:46 keyring_linux -- common/autotest_common.sh@978 -- # wait 26285 00:39:33.836 00:39:33.836 real 0m5.178s 00:39:33.836 user 0m9.539s 00:39:33.836 sys 0m1.410s 00:39:33.836 19:35:46 keyring_linux -- common/autotest_common.sh@1130 -- # xtrace_disable 00:39:33.836 19:35:46 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:39:33.836 ************************************ 00:39:33.836 END TEST keyring_linux 00:39:33.836 ************************************ 00:39:33.836 19:35:46 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:39:33.836 19:35:46 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:39:33.836 19:35:46 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:39:33.836 19:35:46 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:39:33.836 19:35:46 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:39:33.836 19:35:46 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:39:33.836 19:35:46 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:39:33.836 19:35:46 -- common/autotest_common.sh@726 -- # xtrace_disable 00:39:33.836 19:35:46 -- common/autotest_common.sh@10 -- # set +x 00:39:33.836 19:35:46 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:39:33.836 19:35:46 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:39:33.836 19:35:46 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:39:33.836 19:35:46 -- common/autotest_common.sh@10 -- # set +x 00:39:41.981 INFO: APP EXITING 00:39:41.981 INFO: killing all VMs 00:39:41.981 INFO: killing vhost app 00:39:41.981 INFO: EXIT DONE 00:39:45.283 0000:80:01.6 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:80:01.7 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:80:01.4 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:80:01.5 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:80:01.2 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:80:01.3 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:80:01.0 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:80:01.1 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:65:00.0 (144d a80a): Already using the nvme driver 00:39:45.283 0000:00:01.6 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:00:01.7 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:00:01.4 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:00:01.5 (8086 0b00): Already using the ioatdma driver 00:39:45.283 0000:00:01.2 (8086 0b00): Already using the ioatdma driver 00:39:45.543 0000:00:01.3 (8086 0b00): Already using the ioatdma driver 00:39:45.543 0000:00:01.0 (8086 0b00): Already using the ioatdma driver 00:39:45.543 0000:00:01.1 (8086 0b00): Already using the ioatdma driver 00:39:49.748 Cleaning 00:39:49.748 Removing: /var/run/dpdk/spdk0/config 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-0 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-1 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-2 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-3 00:39:49.748 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:39:49.748 Removing: /var/run/dpdk/spdk0/hugepage_info 00:39:49.748 Removing: /var/run/dpdk/spdk1/config 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-0 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-1 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-2 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-3 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-0 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-1 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-2 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-3 00:39:49.748 Removing: /var/run/dpdk/spdk1/fbarray_memzone 00:39:49.748 Removing: /var/run/dpdk/spdk1/hugepage_info 00:39:49.748 Removing: /var/run/dpdk/spdk2/config 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-0 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-1 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-2 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-3 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-0 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-1 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-2 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-3 00:39:49.748 Removing: /var/run/dpdk/spdk2/fbarray_memzone 00:39:49.748 Removing: /var/run/dpdk/spdk2/hugepage_info 00:39:49.748 Removing: /var/run/dpdk/spdk3/config 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-0 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-1 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-2 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-3 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-0 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-1 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-2 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-3 00:39:49.748 Removing: /var/run/dpdk/spdk3/fbarray_memzone 00:39:49.748 Removing: /var/run/dpdk/spdk3/hugepage_info 00:39:49.748 Removing: /var/run/dpdk/spdk4/config 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-0 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-1 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-2 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-3 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-0 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-1 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-2 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-3 00:39:49.748 Removing: /var/run/dpdk/spdk4/fbarray_memzone 00:39:49.748 Removing: /var/run/dpdk/spdk4/hugepage_info 00:39:49.748 Removing: /dev/shm/bdev_svc_trace.1 00:39:49.748 Removing: /dev/shm/nvmf_trace.0 00:39:49.748 Removing: /dev/shm/spdk_tgt_trace.pid3600186 00:39:49.748 Removing: /var/run/dpdk/spdk0 00:39:49.748 Removing: /var/run/dpdk/spdk1 00:39:49.748 Removing: /var/run/dpdk/spdk2 00:39:49.748 Removing: /var/run/dpdk/spdk3 00:39:49.748 Removing: /var/run/dpdk/spdk4 00:39:49.748 Removing: /var/run/dpdk/spdk_pid13493 00:39:49.748 Removing: /var/run/dpdk/spdk_pid14109 00:39:49.748 Removing: /var/run/dpdk/spdk_pid14654 00:39:49.748 Removing: /var/run/dpdk/spdk_pid17661 00:39:49.748 Removing: /var/run/dpdk/spdk_pid18240 00:39:49.748 Removing: /var/run/dpdk/spdk_pid18911 00:39:49.748 Removing: /var/run/dpdk/spdk_pid2238 00:39:49.748 Removing: /var/run/dpdk/spdk_pid24021 00:39:49.748 Removing: /var/run/dpdk/spdk_pid24042 00:39:49.748 Removing: /var/run/dpdk/spdk_pid25842 00:39:49.748 Removing: /var/run/dpdk/spdk_pid26285 00:39:49.748 Removing: /var/run/dpdk/spdk_pid26621 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3598003 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3600186 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3600733 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3601916 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3602115 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3603420 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3603514 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3603965 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3605095 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3605573 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3605968 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3606361 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3606787 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3607183 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3607543 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3607726 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3607984 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3609350 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3612616 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3612977 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3613344 00:39:49.748 Removing: /var/run/dpdk/spdk_pid3613461 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3613939 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3614058 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3614429 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3614524 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3614811 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3615142 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3615232 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3615518 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3615971 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3616321 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3616725 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3621947 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3627686 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3640172 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3641029 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3646892 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3647358 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3653517 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3660997 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3664115 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3677895 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3690036 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3692107 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3693124 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3716415 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3721527 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3783028 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3789954 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3797685 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3806042 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3806120 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3807136 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3808163 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3809219 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3809857 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3810012 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3810237 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3810390 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3810400 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3811405 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3812523 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3813717 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3814666 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3814698 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3815056 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3816496 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3817921 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3828591 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3865907 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3871900 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3873849 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3876337 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3876388 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3876685 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3876761 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3877596 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3879854 00:39:50.010 Removing: /var/run/dpdk/spdk_pid3880975 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3881494 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3884108 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3884850 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3885841 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3891266 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3898463 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3898465 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3898467 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3904412 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3915786 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3920788 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3928689 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3930240 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3931924 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3933651 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3939716 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3945578 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3951230 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3961831 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3961891 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3967671 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3967868 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3968033 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3968667 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3968696 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3974756 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3975438 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3981405 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3984516 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3991601 00:39:50.271 Removing: /var/run/dpdk/spdk_pid3998757 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4009566 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4020003 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4020005 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4045778 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4046460 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4047183 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4048038 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4048979 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4049748 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4050254 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4051062 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4056667 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4057001 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4064727 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4064962 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4072494 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4078189 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4090244 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4090988 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4096649 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4097004 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4102646 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4109972 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4112876 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4126607 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4138154 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4140055 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4141199 00:39:50.271 Removing: /var/run/dpdk/spdk_pid4162356 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4167490 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4170817 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4179249 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4179288 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4185968 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4188166 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4190495 00:39:50.533 Removing: /var/run/dpdk/spdk_pid4191878 00:39:50.533 Removing: /var/run/dpdk/spdk_pid513 00:39:50.533 Clean 00:39:50.533 19:36:03 -- common/autotest_common.sh@1453 -- # return 0 00:39:50.533 19:36:03 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:39:50.533 19:36:03 -- common/autotest_common.sh@732 -- # xtrace_disable 00:39:50.533 19:36:03 -- common/autotest_common.sh@10 -- # set +x 00:39:50.533 19:36:03 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:39:50.533 19:36:03 -- common/autotest_common.sh@732 -- # xtrace_disable 00:39:50.533 19:36:03 -- common/autotest_common.sh@10 -- # set +x 00:39:50.533 19:36:03 -- spdk/autotest.sh@392 -- # chmod a+r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:39:50.533 19:36:03 -- spdk/autotest.sh@394 -- # [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log ]] 00:39:50.533 19:36:03 -- spdk/autotest.sh@394 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log 00:39:50.533 19:36:03 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:39:50.533 19:36:03 -- spdk/autotest.sh@398 -- # hostname 00:39:50.533 19:36:03 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -t spdk-cyp-12 -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info 00:39:50.793 geninfo: WARNING: invalid characters removed from testname! 00:40:17.380 19:36:28 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:40:19.293 19:36:31 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/dpdk/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:40:20.689 19:36:33 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:40:22.600 19:36:34 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/examples/vmd/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:40:23.983 19:36:36 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:40:25.892 19:36:38 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:40:27.275 19:36:39 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:40:27.275 19:36:39 -- spdk/autorun.sh@1 -- $ timing_finish 00:40:27.275 19:36:39 -- common/autotest_common.sh@738 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt ]] 00:40:27.275 19:36:39 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:40:27.275 19:36:39 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:40:27.275 19:36:39 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:40:27.275 + [[ -n 3512425 ]] 00:40:27.275 + sudo kill 3512425 00:40:27.285 [Pipeline] } 00:40:27.300 [Pipeline] // stage 00:40:27.305 [Pipeline] } 00:40:27.318 [Pipeline] // timeout 00:40:27.323 [Pipeline] } 00:40:27.337 [Pipeline] // catchError 00:40:27.342 [Pipeline] } 00:40:27.356 [Pipeline] // wrap 00:40:27.362 [Pipeline] } 00:40:27.374 [Pipeline] // catchError 00:40:27.382 [Pipeline] stage 00:40:27.384 [Pipeline] { (Epilogue) 00:40:27.397 [Pipeline] catchError 00:40:27.398 [Pipeline] { 00:40:27.409 [Pipeline] echo 00:40:27.410 Cleanup processes 00:40:27.415 [Pipeline] sh 00:40:27.697 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:40:27.697 40571 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:40:27.711 [Pipeline] sh 00:40:27.998 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:40:27.998 ++ grep -v 'sudo pgrep' 00:40:27.998 ++ awk '{print $1}' 00:40:27.998 + sudo kill -9 00:40:27.998 + true 00:40:28.012 [Pipeline] sh 00:40:28.300 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:40:40.544 [Pipeline] sh 00:40:40.826 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:40:40.826 Artifacts sizes are good 00:40:40.841 [Pipeline] archiveArtifacts 00:40:40.849 Archiving artifacts 00:40:41.022 [Pipeline] sh 00:40:41.378 + sudo chown -R sys_sgci: /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:40:41.394 [Pipeline] cleanWs 00:40:41.405 [WS-CLEANUP] Deleting project workspace... 00:40:41.405 [WS-CLEANUP] Deferred wipeout is used... 00:40:41.413 [WS-CLEANUP] done 00:40:41.415 [Pipeline] } 00:40:41.433 [Pipeline] // catchError 00:40:41.445 [Pipeline] sh 00:40:41.736 + logger -p user.info -t JENKINS-CI 00:40:41.746 [Pipeline] } 00:40:41.760 [Pipeline] // stage 00:40:41.766 [Pipeline] } 00:40:41.783 [Pipeline] // node 00:40:41.788 [Pipeline] End of Pipeline 00:40:41.826 Finished: SUCCESS